var/home/core/zuul-output/0000755000175000017500000000000015110526306014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110550130015460 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005113761515110550120017674 0ustar rootrootNov 23 06:49:43 crc systemd[1]: Starting Kubernetes Kubelet... Nov 23 06:49:43 crc restorecon[4818]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:43 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:49:44 crc restorecon[4818]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 23 06:49:45 crc kubenswrapper[4906]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:49:45 crc kubenswrapper[4906]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 23 06:49:45 crc kubenswrapper[4906]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:49:45 crc kubenswrapper[4906]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:49:45 crc kubenswrapper[4906]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 23 06:49:45 crc kubenswrapper[4906]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.054392 4906 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060830 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060867 4906 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060880 4906 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060892 4906 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060904 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060914 4906 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060922 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060930 4906 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060939 4906 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060948 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060955 4906 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060963 4906 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060971 4906 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060978 4906 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060986 4906 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.060994 4906 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061002 4906 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061010 4906 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061018 4906 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061025 4906 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061034 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061041 4906 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061049 4906 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061056 4906 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061064 4906 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061072 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061080 4906 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061105 4906 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061114 4906 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061125 4906 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061135 4906 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061148 4906 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061161 4906 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061174 4906 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061185 4906 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061197 4906 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061209 4906 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061221 4906 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061231 4906 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061243 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061253 4906 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061264 4906 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061274 4906 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061286 4906 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061296 4906 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061306 4906 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061317 4906 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061327 4906 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061340 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061351 4906 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061361 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061372 4906 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061381 4906 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061391 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061400 4906 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061409 4906 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061419 4906 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061429 4906 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061439 4906 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061448 4906 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061457 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061470 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061480 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061491 4906 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061501 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061514 4906 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061526 4906 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061539 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061550 4906 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061561 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.061572 4906 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.062868 4906 flags.go:64] FLAG: --address="0.0.0.0" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.062943 4906 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.062976 4906 flags.go:64] FLAG: --anonymous-auth="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.062991 4906 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063006 4906 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063016 4906 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063031 4906 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063048 4906 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063058 4906 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063068 4906 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063079 4906 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063089 4906 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063098 4906 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063108 4906 flags.go:64] FLAG: --cgroup-root="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063117 4906 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063127 4906 flags.go:64] FLAG: --client-ca-file="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063135 4906 flags.go:64] FLAG: --cloud-config="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063144 4906 flags.go:64] FLAG: --cloud-provider="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063154 4906 flags.go:64] FLAG: --cluster-dns="[]" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063165 4906 flags.go:64] FLAG: --cluster-domain="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063174 4906 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063184 4906 flags.go:64] FLAG: --config-dir="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063193 4906 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063203 4906 flags.go:64] FLAG: --container-log-max-files="5" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063223 4906 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063233 4906 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063243 4906 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063252 4906 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063261 4906 flags.go:64] FLAG: --contention-profiling="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063270 4906 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063280 4906 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063290 4906 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063298 4906 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063315 4906 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063326 4906 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063336 4906 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063348 4906 flags.go:64] FLAG: --enable-load-reader="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063359 4906 flags.go:64] FLAG: --enable-server="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063369 4906 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063383 4906 flags.go:64] FLAG: --event-burst="100" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063394 4906 flags.go:64] FLAG: --event-qps="50" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063403 4906 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063412 4906 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063422 4906 flags.go:64] FLAG: --eviction-hard="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063435 4906 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063444 4906 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063454 4906 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063463 4906 flags.go:64] FLAG: --eviction-soft="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063472 4906 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063481 4906 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063491 4906 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063500 4906 flags.go:64] FLAG: --experimental-mounter-path="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063509 4906 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063518 4906 flags.go:64] FLAG: --fail-swap-on="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063527 4906 flags.go:64] FLAG: --feature-gates="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063539 4906 flags.go:64] FLAG: --file-check-frequency="20s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063549 4906 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063558 4906 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063568 4906 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063578 4906 flags.go:64] FLAG: --healthz-port="10248" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063587 4906 flags.go:64] FLAG: --help="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063597 4906 flags.go:64] FLAG: --hostname-override="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063606 4906 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063624 4906 flags.go:64] FLAG: --http-check-frequency="20s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063633 4906 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063643 4906 flags.go:64] FLAG: --image-credential-provider-config="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063652 4906 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063663 4906 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063672 4906 flags.go:64] FLAG: --image-service-endpoint="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063714 4906 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063724 4906 flags.go:64] FLAG: --kube-api-burst="100" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063734 4906 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063745 4906 flags.go:64] FLAG: --kube-api-qps="50" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063755 4906 flags.go:64] FLAG: --kube-reserved="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063764 4906 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063773 4906 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063782 4906 flags.go:64] FLAG: --kubelet-cgroups="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063791 4906 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063801 4906 flags.go:64] FLAG: --lock-file="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063809 4906 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063819 4906 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063828 4906 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063846 4906 flags.go:64] FLAG: --log-json-split-stream="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063855 4906 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063864 4906 flags.go:64] FLAG: --log-text-split-stream="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063873 4906 flags.go:64] FLAG: --logging-format="text" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063882 4906 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063892 4906 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063901 4906 flags.go:64] FLAG: --manifest-url="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063910 4906 flags.go:64] FLAG: --manifest-url-header="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063925 4906 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063934 4906 flags.go:64] FLAG: --max-open-files="1000000" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063946 4906 flags.go:64] FLAG: --max-pods="110" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063955 4906 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063965 4906 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063978 4906 flags.go:64] FLAG: --memory-manager-policy="None" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063987 4906 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.063998 4906 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064008 4906 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064017 4906 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064051 4906 flags.go:64] FLAG: --node-status-max-images="50" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064061 4906 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064070 4906 flags.go:64] FLAG: --oom-score-adj="-999" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064080 4906 flags.go:64] FLAG: --pod-cidr="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064089 4906 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064104 4906 flags.go:64] FLAG: --pod-manifest-path="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064112 4906 flags.go:64] FLAG: --pod-max-pids="-1" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064122 4906 flags.go:64] FLAG: --pods-per-core="0" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064133 4906 flags.go:64] FLAG: --port="10250" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064143 4906 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064152 4906 flags.go:64] FLAG: --provider-id="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064161 4906 flags.go:64] FLAG: --qos-reserved="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064170 4906 flags.go:64] FLAG: --read-only-port="10255" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064179 4906 flags.go:64] FLAG: --register-node="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064189 4906 flags.go:64] FLAG: --register-schedulable="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064198 4906 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064221 4906 flags.go:64] FLAG: --registry-burst="10" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064230 4906 flags.go:64] FLAG: --registry-qps="5" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064239 4906 flags.go:64] FLAG: --reserved-cpus="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064248 4906 flags.go:64] FLAG: --reserved-memory="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064260 4906 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064269 4906 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064279 4906 flags.go:64] FLAG: --rotate-certificates="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064289 4906 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064298 4906 flags.go:64] FLAG: --runonce="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064308 4906 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064317 4906 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064332 4906 flags.go:64] FLAG: --seccomp-default="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064342 4906 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064351 4906 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064370 4906 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064379 4906 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064390 4906 flags.go:64] FLAG: --storage-driver-password="root" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064400 4906 flags.go:64] FLAG: --storage-driver-secure="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064409 4906 flags.go:64] FLAG: --storage-driver-table="stats" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064419 4906 flags.go:64] FLAG: --storage-driver-user="root" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064429 4906 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064439 4906 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064449 4906 flags.go:64] FLAG: --system-cgroups="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064458 4906 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064474 4906 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064483 4906 flags.go:64] FLAG: --tls-cert-file="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064492 4906 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064503 4906 flags.go:64] FLAG: --tls-min-version="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064514 4906 flags.go:64] FLAG: --tls-private-key-file="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064523 4906 flags.go:64] FLAG: --topology-manager-policy="none" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064532 4906 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064541 4906 flags.go:64] FLAG: --topology-manager-scope="container" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064550 4906 flags.go:64] FLAG: --v="2" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064565 4906 flags.go:64] FLAG: --version="false" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064578 4906 flags.go:64] FLAG: --vmodule="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064590 4906 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.064600 4906 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.064934 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.064946 4906 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.064958 4906 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.064967 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.064976 4906 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.064984 4906 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.064995 4906 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065003 4906 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065011 4906 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065021 4906 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065029 4906 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065038 4906 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065046 4906 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065055 4906 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065062 4906 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065070 4906 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065078 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065086 4906 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065094 4906 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065101 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065109 4906 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065117 4906 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065125 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065133 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065140 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065148 4906 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065155 4906 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065164 4906 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065173 4906 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065180 4906 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065188 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065195 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065203 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065212 4906 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065220 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065228 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065236 4906 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065243 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065252 4906 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065260 4906 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065272 4906 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065284 4906 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065294 4906 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065304 4906 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065313 4906 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065323 4906 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065333 4906 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065343 4906 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065351 4906 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065360 4906 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065367 4906 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065375 4906 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065382 4906 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065390 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065398 4906 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065405 4906 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065415 4906 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065425 4906 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065434 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065442 4906 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065450 4906 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065457 4906 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065466 4906 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065475 4906 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065484 4906 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065492 4906 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065499 4906 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065508 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065517 4906 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065525 4906 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.065534 4906 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.065560 4906 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.079695 4906 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.079750 4906 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079841 4906 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079853 4906 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079859 4906 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079864 4906 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079869 4906 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079874 4906 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079879 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079884 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079891 4906 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079896 4906 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079900 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079905 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079910 4906 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079915 4906 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079919 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079924 4906 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079929 4906 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079933 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079938 4906 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079942 4906 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079946 4906 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079951 4906 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079957 4906 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079964 4906 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079970 4906 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079975 4906 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079980 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079985 4906 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079990 4906 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.079995 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080002 4906 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080007 4906 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080011 4906 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080017 4906 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080028 4906 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080033 4906 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080037 4906 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080042 4906 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080049 4906 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080056 4906 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080062 4906 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080068 4906 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080075 4906 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080080 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080086 4906 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080093 4906 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080098 4906 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080103 4906 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080108 4906 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080113 4906 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080117 4906 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080121 4906 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080126 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080131 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080136 4906 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080141 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080146 4906 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080151 4906 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080156 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080161 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080166 4906 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080172 4906 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080177 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080182 4906 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080186 4906 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080191 4906 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080196 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080201 4906 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080206 4906 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080210 4906 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080216 4906 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.080225 4906 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080382 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080391 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080397 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080402 4906 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080407 4906 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080412 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080417 4906 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080422 4906 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080427 4906 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080433 4906 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080438 4906 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080443 4906 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080449 4906 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080456 4906 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080462 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080468 4906 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080473 4906 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080478 4906 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080483 4906 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080488 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080493 4906 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080497 4906 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080503 4906 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080507 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080512 4906 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080517 4906 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080522 4906 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080528 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080533 4906 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080538 4906 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080544 4906 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080548 4906 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080553 4906 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080559 4906 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080564 4906 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080569 4906 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080573 4906 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080578 4906 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080582 4906 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080587 4906 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080591 4906 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080596 4906 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080600 4906 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080605 4906 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080609 4906 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080614 4906 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080619 4906 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080624 4906 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080629 4906 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080633 4906 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080638 4906 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080643 4906 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080648 4906 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080654 4906 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080658 4906 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080664 4906 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080670 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080698 4906 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080706 4906 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080712 4906 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080716 4906 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080721 4906 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080726 4906 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080730 4906 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080735 4906 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080740 4906 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080745 4906 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080750 4906 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080754 4906 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080760 4906 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.080769 4906 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.080779 4906 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.081885 4906 server.go:940] "Client rotation is on, will bootstrap in background" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.087170 4906 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.087280 4906 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.089052 4906 server.go:997] "Starting client certificate rotation" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.089082 4906 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.090666 4906 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-30 08:03:51.344674348 +0000 UTC Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.090883 4906 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 889h14m6.253800203s for next certificate rotation Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.129012 4906 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.133637 4906 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.156475 4906 log.go:25] "Validated CRI v1 runtime API" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.205003 4906 log.go:25] "Validated CRI v1 image API" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.207882 4906 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.215393 4906 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-23-06-40-50-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.215492 4906 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.244557 4906 manager.go:217] Machine: {Timestamp:2025-11-23 06:49:45.241014811 +0000 UTC m=+0.754406184 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:9f63f9ce-8df2-4d0e-becf-5452e97f2c1a BootID:9217c3d2-062e-44c4-9092-623fa669de19 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f5:a7:89 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f5:a7:89 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:4f:71:eb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1f:33:82 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:23:02:5c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e6:8a:03 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:23:66:e0 Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:dd:70:dd Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ae:24:d6:c4:3c:65 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a6:36:70:57:d3:0b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.245032 4906 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.245263 4906 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.246627 4906 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.246998 4906 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.247057 4906 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.247438 4906 topology_manager.go:138] "Creating topology manager with none policy" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.247457 4906 container_manager_linux.go:303] "Creating device plugin manager" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.248190 4906 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.248244 4906 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.248631 4906 state_mem.go:36] "Initialized new in-memory state store" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.248831 4906 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.256675 4906 kubelet.go:418] "Attempting to sync node with API server" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.256768 4906 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.256842 4906 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.256875 4906 kubelet.go:324] "Adding apiserver pod source" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.256902 4906 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.263165 4906 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.264602 4906 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.266975 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.267181 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.267168 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.267380 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.268721 4906 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.271858 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.271902 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.271932 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.271947 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.271973 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.271988 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.272003 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.272026 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.272044 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.272061 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.272082 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.272105 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.273286 4906 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.274289 4906 server.go:1280] "Started kubelet" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.275291 4906 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.276433 4906 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 23 06:49:45 crc systemd[1]: Started Kubernetes Kubelet. Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.276438 4906 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.282837 4906 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.287714 4906 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.287770 4906 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.288182 4906 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 18:47:51.81023297 +0000 UTC Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.288379 4906 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.288395 4906 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.288470 4906 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.289288 4906 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.290484 4906 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="200ms" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.290284 4906 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.227:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a900bff7a4beb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-23 06:49:45.274215403 +0000 UTC m=+0.787606746,LastTimestamp:2025-11-23 06:49:45.274215403 +0000 UTC m=+0.787606746,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.296862 4906 server.go:460] "Adding debug handlers to kubelet server" Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.298024 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.298192 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.298279 4906 factory.go:153] Registering CRI-O factory Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.298327 4906 factory.go:221] Registration of the crio container factory successfully Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.298460 4906 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.298478 4906 factory.go:55] Registering systemd factory Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.298492 4906 factory.go:221] Registration of the systemd container factory successfully Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.298530 4906 factory.go:103] Registering Raw factory Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.298590 4906 manager.go:1196] Started watching for new ooms in manager Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.299674 4906 manager.go:319] Starting recovery of all containers Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.314314 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.314618 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.314822 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.314981 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.315106 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.315228 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.315348 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.315491 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.315617 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.315792 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.315969 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.316115 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.316252 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.316381 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.316503 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.316705 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.316862 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.317028 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.317199 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.317355 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.317492 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.317632 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.317813 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.317938 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.318118 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.318262 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.318425 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.318570 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.318738 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.318875 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.318994 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.319162 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.319287 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.319428 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.319553 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.319710 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.319841 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.319969 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.320109 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.320236 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.320354 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.320472 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.320594 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.320744 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.321590 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.322081 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.322109 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.322288 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.322346 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.322394 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.322576 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.322604 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.329258 4906 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.329393 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.329453 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.329492 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.329561 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.329598 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.329625 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.331630 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.331795 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.331818 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.331851 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.331872 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332512 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332541 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332565 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332593 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332698 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332722 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332749 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332765 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332789 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332807 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332820 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332840 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332855 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332870 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332895 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332912 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.332935 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333003 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333018 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333037 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333051 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333070 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333084 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333116 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333134 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333151 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333167 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333182 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333195 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333212 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333228 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333246 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333259 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333272 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333295 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333311 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333324 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333342 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333363 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333380 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333397 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333427 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333452 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333473 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333491 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333509 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333531 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333549 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333563 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333579 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333600 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333615 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333631 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333648 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333696 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333806 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333822 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333839 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333851 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333864 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333882 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333894 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333913 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333925 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333938 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333952 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333965 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333978 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.333991 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334004 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334019 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334031 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334059 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334071 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334084 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334098 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334112 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334124 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334140 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334154 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334169 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334182 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334196 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334209 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334222 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334236 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334248 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334262 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334276 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334288 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334305 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334318 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334335 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334351 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334365 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334383 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334395 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334406 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334421 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334433 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334447 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334459 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334473 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334490 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334504 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334518 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334535 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334549 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334570 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334585 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334603 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334634 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334653 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334690 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334709 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334725 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334750 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334766 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.334789 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.335898 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.335951 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.335964 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.335981 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.335994 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336008 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336021 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336033 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336044 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336065 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336083 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336095 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336108 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336150 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336162 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336174 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336186 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336200 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336214 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336227 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336239 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336251 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336263 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336274 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336288 4906 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336298 4906 reconstruct.go:97] "Volume reconstruction finished" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.336307 4906 reconciler.go:26] "Reconciler: start to sync state" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.342069 4906 manager.go:324] Recovery completed Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.353322 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.353706 4906 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.355166 4906 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.355211 4906 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.355241 4906 kubelet.go:2335] "Starting kubelet main sync loop" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.355351 4906 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.355367 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.355414 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.355428 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: W1123 06:49:45.356835 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.356914 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.357864 4906 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.357888 4906 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.357911 4906 state_mem.go:36] "Initialized new in-memory state store" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.382873 4906 policy_none.go:49] "None policy: Start" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.384500 4906 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.384566 4906 state_mem.go:35] "Initializing new in-memory state store" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.389424 4906 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.455371 4906 manager.go:334] "Starting Device Plugin manager" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.455443 4906 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.455463 4906 server.go:79] "Starting device plugin registration server" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.455586 4906 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.456147 4906 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.456180 4906 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.456898 4906 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.457333 4906 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.457360 4906 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.469413 4906 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.491663 4906 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="400ms" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.557259 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.559204 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.559278 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.559300 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.559351 4906 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.560243 4906 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.656765 4906 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.657377 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.659985 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.660061 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.660091 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.660362 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.660533 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.660595 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662150 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662204 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662221 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662244 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662251 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662257 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662522 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662663 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.662732 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664148 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664203 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664215 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664264 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664275 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664460 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664762 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.664855 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.665451 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.665477 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.665489 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.665656 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.665829 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.665875 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.667116 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.667180 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.667205 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.668553 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.668617 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.668647 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.669206 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.669307 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.669312 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.669403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.669446 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.673823 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.673865 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.673877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.741858 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.741968 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742008 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742049 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742221 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742256 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742305 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742348 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742451 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742536 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742574 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742612 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742664 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742718 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.742749 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.760902 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.762838 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.762906 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.762933 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.762991 4906 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.763719 4906 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844460 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844547 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844591 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844624 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844658 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844724 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844771 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844788 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844900 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844887 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844967 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844964 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845008 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845025 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844815 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845101 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.844970 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845155 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845171 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845189 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845218 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845224 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845250 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845254 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845284 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845313 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845316 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845349 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845287 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.845504 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:45 crc kubenswrapper[4906]: E1123 06:49:45.893563 4906 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="800ms" Nov 23 06:49:45 crc kubenswrapper[4906]: I1123 06:49:45.993655 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.000283 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.023524 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.039055 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.046010 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.056299 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f6ba5a4530eec0c4ea05fa8e1167a477b45ef98799e76d5ef9ceb8c83230840f WatchSource:0}: Error finding container f6ba5a4530eec0c4ea05fa8e1167a477b45ef98799e76d5ef9ceb8c83230840f: Status 404 returned error can't find the container with id f6ba5a4530eec0c4ea05fa8e1167a477b45ef98799e76d5ef9ceb8c83230840f Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.057908 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ea866b61d16e71431a5dd0b287d0540bf082c29e16da99349e990526067f8152 WatchSource:0}: Error finding container ea866b61d16e71431a5dd0b287d0540bf082c29e16da99349e990526067f8152: Status 404 returned error can't find the container with id ea866b61d16e71431a5dd0b287d0540bf082c29e16da99349e990526067f8152 Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.064601 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3c840ae3a058f37bbeb5472d49d7e7ef9ded12f9356e87d87d4d4a1c8cb1a91c WatchSource:0}: Error finding container 3c840ae3a058f37bbeb5472d49d7e7ef9ded12f9356e87d87d4d4a1c8cb1a91c: Status 404 returned error can't find the container with id 3c840ae3a058f37bbeb5472d49d7e7ef9ded12f9356e87d87d4d4a1c8cb1a91c Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.076878 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-df536dc457052b032da7ea88b5f0b66d7b9cdb55c1bea41828ea79de1d139ba3 WatchSource:0}: Error finding container df536dc457052b032da7ea88b5f0b66d7b9cdb55c1bea41828ea79de1d139ba3: Status 404 returned error can't find the container with id df536dc457052b032da7ea88b5f0b66d7b9cdb55c1bea41828ea79de1d139ba3 Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.079308 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-72058d4086dcf064954b8df18d434d57523042e5080414a177b68af52caf725f WatchSource:0}: Error finding container 72058d4086dcf064954b8df18d434d57523042e5080414a177b68af52caf725f: Status 404 returned error can't find the container with id 72058d4086dcf064954b8df18d434d57523042e5080414a177b68af52caf725f Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.165245 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.167282 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.167351 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.167367 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.167412 4906 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:49:46 crc kubenswrapper[4906]: E1123 06:49:46.168189 4906 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.276505 4906 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.288564 4906 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 19:27:17.704020474 +0000 UTC Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.288628 4906 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 996h37m31.415394908s for next certificate rotation Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.362093 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"72058d4086dcf064954b8df18d434d57523042e5080414a177b68af52caf725f"} Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.363605 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"df536dc457052b032da7ea88b5f0b66d7b9cdb55c1bea41828ea79de1d139ba3"} Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.365905 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:46 crc kubenswrapper[4906]: E1123 06:49:46.366016 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.366650 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3c840ae3a058f37bbeb5472d49d7e7ef9ded12f9356e87d87d4d4a1c8cb1a91c"} Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.367960 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ea866b61d16e71431a5dd0b287d0540bf082c29e16da99349e990526067f8152"} Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.369006 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f6ba5a4530eec0c4ea05fa8e1167a477b45ef98799e76d5ef9ceb8c83230840f"} Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.606838 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:46 crc kubenswrapper[4906]: E1123 06:49:46.606942 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.676096 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:46 crc kubenswrapper[4906]: E1123 06:49:46.676417 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:46 crc kubenswrapper[4906]: E1123 06:49:46.694815 4906 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="1.6s" Nov 23 06:49:46 crc kubenswrapper[4906]: W1123 06:49:46.712980 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:46 crc kubenswrapper[4906]: E1123 06:49:46.713483 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.968981 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.970501 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.970553 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.970566 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:46 crc kubenswrapper[4906]: I1123 06:49:46.970604 4906 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:49:46 crc kubenswrapper[4906]: E1123 06:49:46.971260 4906 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.277132 4906 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.374596 4906 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359" exitCode=0 Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.374727 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359"} Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.374911 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.376591 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.376633 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93"} Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.376659 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.376710 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.376596 4906 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93" exitCode=0 Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.377428 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.379317 4906 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5" exitCode=0 Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.379428 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.379437 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5"} Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.379666 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.380018 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.380069 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.380093 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.380611 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.380638 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.380652 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.381405 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.381439 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.381452 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.384114 4906 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b" exitCode=0 Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.384174 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b"} Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.384248 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.385563 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.385661 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.385695 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.391229 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853"} Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.391293 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0"} Nov 23 06:49:47 crc kubenswrapper[4906]: I1123 06:49:47.391321 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.276853 4906 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:48 crc kubenswrapper[4906]: E1123 06:49:48.295982 4906 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.227:6443: connect: connection refused" interval="3.2s" Nov 23 06:49:48 crc kubenswrapper[4906]: W1123 06:49:48.307154 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:48 crc kubenswrapper[4906]: E1123 06:49:48.307273 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.402867 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0128e3b7c58d98ebc49b75ecad88f2641115aff47aaef29a2a6bba6795c2b641"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.403024 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.406074 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.406289 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.406313 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.409802 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.409833 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.409847 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.410019 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.411724 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.411755 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.411766 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.416540 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.416652 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.418337 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.418355 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.418363 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.424888 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.424950 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.424967 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.424982 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.435242 4906 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6" exitCode=0 Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.435322 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6"} Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.435512 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.436718 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.436763 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.436779 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:48 crc kubenswrapper[4906]: W1123 06:49:48.443732 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:48 crc kubenswrapper[4906]: E1123 06:49:48.443831 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.572171 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.573587 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.573633 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.573645 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:48 crc kubenswrapper[4906]: I1123 06:49:48.573693 4906 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:49:48 crc kubenswrapper[4906]: E1123 06:49:48.574259 4906 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.227:6443: connect: connection refused" node="crc" Nov 23 06:49:48 crc kubenswrapper[4906]: W1123 06:49:48.590201 4906 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.227:6443: connect: connection refused Nov 23 06:49:48 crc kubenswrapper[4906]: E1123 06:49:48.590291 4906 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.227:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.379926 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.389960 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.443801 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4"} Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.444052 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.445820 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.445870 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.445889 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.449952 4906 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2" exitCode=0 Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.450110 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.450012 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2"} Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.450244 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.450373 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.451307 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.451354 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.451374 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.451875 4906 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.451963 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.454099 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.454167 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.454190 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.454293 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.454338 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.454361 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.455284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.455317 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.455333 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:49 crc kubenswrapper[4906]: I1123 06:49:49.930040 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.058759 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.458285 4906 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.458316 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.458366 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.458411 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.458301 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23"} Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.458554 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345"} Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.458579 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94"} Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.460198 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.460276 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.460300 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.461602 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.461650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.461695 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.461606 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.462167 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:50 crc kubenswrapper[4906]: I1123 06:49:50.462192 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.232343 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.277723 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.470617 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2"} Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.470737 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9"} Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.470756 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.470864 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.470955 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.472877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.472935 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.473007 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.473034 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.473074 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.473096 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.473109 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.473120 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.473137 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.774760 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.777514 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.777592 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.777611 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:51 crc kubenswrapper[4906]: I1123 06:49:51.777655 4906 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.474226 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.474256 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.476095 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.476159 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.476179 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.476441 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.476490 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.476510 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.686476 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.686819 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.689096 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.689157 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:52 crc kubenswrapper[4906]: I1123 06:49:52.689171 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:53 crc kubenswrapper[4906]: I1123 06:49:53.222925 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 23 06:49:53 crc kubenswrapper[4906]: I1123 06:49:53.477175 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:53 crc kubenswrapper[4906]: I1123 06:49:53.478551 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:53 crc kubenswrapper[4906]: I1123 06:49:53.478626 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:53 crc kubenswrapper[4906]: I1123 06:49:53.478647 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:54 crc kubenswrapper[4906]: I1123 06:49:54.394883 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:49:54 crc kubenswrapper[4906]: I1123 06:49:54.395188 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:54 crc kubenswrapper[4906]: I1123 06:49:54.396988 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:54 crc kubenswrapper[4906]: I1123 06:49:54.397036 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:54 crc kubenswrapper[4906]: I1123 06:49:54.397057 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:55 crc kubenswrapper[4906]: E1123 06:49:55.469539 4906 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 23 06:49:55 crc kubenswrapper[4906]: I1123 06:49:55.687737 4906 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 23 06:49:55 crc kubenswrapper[4906]: I1123 06:49:55.688043 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.163777 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.164054 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.165888 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.165962 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.165980 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.843946 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.844258 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.846028 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.846081 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.846094 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:57 crc kubenswrapper[4906]: I1123 06:49:57.849444 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:49:58 crc kubenswrapper[4906]: I1123 06:49:58.491569 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:49:58 crc kubenswrapper[4906]: I1123 06:49:58.492662 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:49:58 crc kubenswrapper[4906]: I1123 06:49:58.492747 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:49:58 crc kubenswrapper[4906]: I1123 06:49:58.492765 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:49:59 crc kubenswrapper[4906]: I1123 06:49:59.277131 4906 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 23 06:49:59 crc kubenswrapper[4906]: I1123 06:49:59.343096 4906 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 23 06:49:59 crc kubenswrapper[4906]: I1123 06:49:59.343160 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 23 06:49:59 crc kubenswrapper[4906]: I1123 06:49:59.351006 4906 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 23 06:49:59 crc kubenswrapper[4906]: I1123 06:49:59.351083 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.285544 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.285810 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.287132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.287177 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.287198 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.292369 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.500090 4906 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.500466 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.501812 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.501862 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:01 crc kubenswrapper[4906]: I1123 06:50:01.501879 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:04 crc kubenswrapper[4906]: E1123 06:50:04.343895 4906 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.348665 4906 trace.go:236] Trace[1479394283]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:49:49.339) (total time: 15009ms): Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[1479394283]: ---"Objects listed" error: 15009ms (06:50:04.348) Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[1479394283]: [15.009387937s] [15.009387937s] END Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.348725 4906 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.350835 4906 trace.go:236] Trace[655596099]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:49:53.678) (total time: 10672ms): Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[655596099]: ---"Objects listed" error: 10672ms (06:50:04.350) Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[655596099]: [10.672123763s] [10.672123763s] END Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.350958 4906 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 23 06:50:04 crc kubenswrapper[4906]: E1123 06:50:04.351878 4906 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.352431 4906 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.352565 4906 trace.go:236] Trace[470461301]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:49:54.124) (total time: 10228ms): Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[470461301]: ---"Objects listed" error: 10228ms (06:50:04.352) Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[470461301]: [10.22839559s] [10.22839559s] END Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.352636 4906 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.354113 4906 trace.go:236] Trace[977669176]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:49:52.138) (total time: 12215ms): Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[977669176]: ---"Objects listed" error: 12215ms (06:50:04.353) Nov 23 06:50:04 crc kubenswrapper[4906]: Trace[977669176]: [12.215840915s] [12.215840915s] END Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.354133 4906 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.386164 4906 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59168->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.386852 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59168->192.168.126.11:17697: read: connection reset by peer" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.386161 4906 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59174->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.386950 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59174->192.168.126.11:17697: read: connection reset by peer" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.387367 4906 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.387439 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.434911 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.447375 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.508273 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.510233 4906 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4" exitCode=255 Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.510281 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4"} Nov 23 06:50:04 crc kubenswrapper[4906]: I1123 06:50:04.614577 4906 scope.go:117] "RemoveContainer" containerID="323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.271501 4906 apiserver.go:52] "Watching apiserver" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.274183 4906 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.274670 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-jv7sg","openshift-kube-apiserver/kube-apiserver-crc","openshift-multus/multus-lm9xt","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-machine-config-operator/machine-config-daemon-2xgvr","openshift-multus/multus-additional-cni-plugins-6mz65","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275137 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.275265 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275193 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.275422 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275136 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275567 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275601 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275780 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275877 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.275933 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.275806 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.276079 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.276154 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.277920 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.278305 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.279161 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.279173 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.280394 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.280700 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.280704 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.280889 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.280955 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.281027 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.281104 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.281130 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.281163 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.281899 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.282949 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.282982 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.284750 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.284878 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.284898 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.284912 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.284952 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.284877 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.284991 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.285350 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.290492 4906 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.296552 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.308314 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.318661 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.328146 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.339670 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.349375 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.359753 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361622 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361714 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361740 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361759 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361775 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361793 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361812 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361828 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361848 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361886 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361909 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361927 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.361979 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362038 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362065 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362089 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362114 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362134 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362152 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362169 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362193 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362193 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362215 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362232 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362249 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362269 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362294 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362319 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362334 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362350 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362370 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362392 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362394 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362414 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362434 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362459 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362484 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362507 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362532 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362555 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362579 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362603 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362627 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362650 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362691 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362718 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362742 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362765 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362787 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362809 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362831 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362854 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362877 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362900 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362924 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362941 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362958 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362978 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362900 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.362996 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363091 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363137 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363182 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363182 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363210 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363244 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363272 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363304 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363345 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363372 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363400 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363426 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363449 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363475 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363498 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363526 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363558 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363632 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363666 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363709 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363733 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363754 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363782 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363805 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363834 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363861 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363871 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363883 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363914 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363972 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364001 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364030 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364059 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364084 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364114 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364139 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364168 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364193 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364218 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364242 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364406 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364441 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364469 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364494 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364520 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364582 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364609 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364634 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364665 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364712 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364737 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364764 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364792 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364817 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364840 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364865 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364892 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364919 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365341 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365381 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365408 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365436 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365461 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365491 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365522 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365548 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365576 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365605 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365630 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365654 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365698 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365722 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365752 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365778 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365804 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365832 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365858 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365886 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365917 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365944 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365979 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366014 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366045 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366075 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366101 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366133 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366172 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366210 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366248 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366287 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366317 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366341 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366365 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366391 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366416 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366439 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366465 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366488 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366521 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366547 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366570 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366594 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366626 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366652 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366738 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366768 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366821 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366859 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366883 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366907 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366932 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366968 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367018 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367046 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367074 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367100 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367124 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367149 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367176 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367200 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367226 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367253 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367282 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367310 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367338 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367368 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367396 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367424 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367456 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367481 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367659 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367736 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367775 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367843 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367886 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367944 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367983 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368011 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368039 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368070 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368150 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-cni-multus\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368205 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368238 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ec5c6569-64d4-4591-bbac-b70ebcca836f-rootfs\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368264 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7hrg\" (UniqueName: \"kubernetes.io/projected/46e886ca-acf1-4bf0-94fd-69dd208d8da9-kube-api-access-w7hrg\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368301 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368329 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec5c6569-64d4-4591-bbac-b70ebcca836f-proxy-tls\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368355 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-cni-bin\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368387 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368413 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cnibin\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368445 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4f75f165-ef56-40be-bfd4-1843de92b356-cni-binary-copy\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368482 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqvhh\" (UniqueName: \"kubernetes.io/projected/c881afa1-c693-450f-912c-cbb4c33bb04f-kube-api-access-nqvhh\") pod \"node-resolver-jv7sg\" (UID: \"c881afa1-c693-450f-912c-cbb4c33bb04f\") " pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368517 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-system-cni-dir\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368559 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368593 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-kubelet\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368635 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368671 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368749 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368796 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368834 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4f75f165-ef56-40be-bfd4-1843de92b356-multus-daemon-config\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368870 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-etc-kubernetes\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370815 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmffv\" (UniqueName: \"kubernetes.io/projected/4f75f165-ef56-40be-bfd4-1843de92b356-kube-api-access-rmffv\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370861 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec5c6569-64d4-4591-bbac-b70ebcca836f-mcd-auth-proxy-config\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370896 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370939 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370974 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-os-release\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371012 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371064 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-cni-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371097 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-socket-dir-parent\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371136 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-hostroot\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371169 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-conf-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371202 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4j66\" (UniqueName: \"kubernetes.io/projected/ec5c6569-64d4-4591-bbac-b70ebcca836f-kube-api-access-d4j66\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371235 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371276 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-k8s-cni-cncf-io\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371307 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-multus-certs\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371342 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c881afa1-c693-450f-912c-cbb4c33bb04f-hosts-file\") pod \"node-resolver-jv7sg\" (UID: \"c881afa1-c693-450f-912c-cbb4c33bb04f\") " pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371384 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371419 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cni-binary-copy\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371443 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-os-release\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371465 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-system-cni-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371490 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-cnibin\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371513 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-netns\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371537 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371568 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371597 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371727 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371749 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371767 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371784 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371800 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.363918 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364411 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364430 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364671 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364561 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364920 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365076 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365124 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365407 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365416 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365453 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.374556 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365469 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.364944 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365865 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.365897 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366074 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366136 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366218 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366254 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366394 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366668 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.366996 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367072 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367329 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367540 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367566 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367689 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367874 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367880 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.367958 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368267 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368290 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368314 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368515 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.368724 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.369059 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.369136 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.369468 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.369668 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.369820 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370510 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370839 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.370856 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371076 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371097 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371114 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371807 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.372501 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.372782 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373007 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373024 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373213 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373262 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373413 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373401 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373418 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373606 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.373711 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.371654 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.374808 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.374841 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.374972 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.375290 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.375338 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.375500 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.375865 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tkkm2"] Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.375994 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.375381 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.376400 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.376645 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.376716 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.376735 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.376804 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.378980 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.379189 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.379235 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.379293 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.379297 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.379339 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.379610 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.379660 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.380389 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.381667 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.381731 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.382027 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.382240 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.382272 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.382318 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.382350 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.382626 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.382916 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.383150 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.383657 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.383751 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.383814 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384024 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384040 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384060 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384069 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384159 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384204 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384418 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384773 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.384809 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.385112 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.385594 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.385714 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.385739 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.385932 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.386121 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.386214 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.386400 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.386537 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.387012 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.387092 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.387275 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.387465 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.387590 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.387668 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.387992 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.388450 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.388731 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.389403 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.389601 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.389647 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.389996 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.390847 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.391476 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.392613 4906 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.392624 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.392916 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:05.892889356 +0000 UTC m=+21.406280659 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.393793 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.394059 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.392788 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.394457 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.394477 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.394556 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:05.89453836 +0000 UTC m=+21.407929663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.394287 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.394399 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.395216 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.395398 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.396007 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.396133 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.396536 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.396761 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.396863 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:05.896833108 +0000 UTC m=+21.410224411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.397081 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.397469 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.397533 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.397981 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.398662 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.398997 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.399050 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.399247 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.399463 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.399497 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.405284 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.405405 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.403156 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.399013 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.400322 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.400672 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.402562 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.402601 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.402846 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.404840 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:50:05.904811682 +0000 UTC m=+21.418202975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.406520 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.406778 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.407139 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.407994 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.408006 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.408135 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.408327 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.409801 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.410605 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.410858 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.410881 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.413833 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.413879 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.413905 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.413925 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.414007 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:05.913986141 +0000 UTC m=+21.427377674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.415922 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.416716 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.417325 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.418127 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.418277 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.418629 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.418767 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.419543 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.420117 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.420440 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.421122 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.421998 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.422964 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.423060 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.423195 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.423295 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.423765 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.425624 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.425797 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.426464 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.426939 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.427060 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.427216 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.427262 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.427311 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.428415 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.428421 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.429830 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.430312 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.430572 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.430862 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.430943 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.433302 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.434544 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.434636 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.436275 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.443063 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.444131 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.449808 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.451383 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.452282 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.458263 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.460848 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.465442 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473042 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473143 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-netd\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473170 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-cni-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473191 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-socket-dir-parent\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473207 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-hostroot\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473224 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-conf-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473242 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4j66\" (UniqueName: \"kubernetes.io/projected/ec5c6569-64d4-4591-bbac-b70ebcca836f-kube-api-access-d4j66\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473258 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-multus-certs\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473276 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-k8s-cni-cncf-io\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473295 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-os-release\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473311 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c881afa1-c693-450f-912c-cbb4c33bb04f-hosts-file\") pod \"node-resolver-jv7sg\" (UID: \"c881afa1-c693-450f-912c-cbb4c33bb04f\") " pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473337 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cni-binary-copy\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473353 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-systemd-units\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473370 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hvsb\" (UniqueName: \"kubernetes.io/projected/92d6f3b5-c353-4412-8f38-989b53ddef2a-kube-api-access-4hvsb\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473385 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-system-cni-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473401 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-cnibin\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473437 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-netns\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473453 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-script-lib\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473468 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-config\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473494 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473510 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-kubelet\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473509 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-cni-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473585 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-hostroot\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473601 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-socket-dir-parent\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473620 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-multus-conf-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473526 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-var-lib-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473671 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473734 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-cni-multus\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473762 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ec5c6569-64d4-4591-bbac-b70ebcca836f-rootfs\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473778 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7hrg\" (UniqueName: \"kubernetes.io/projected/46e886ca-acf1-4bf0-94fd-69dd208d8da9-kube-api-access-w7hrg\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473795 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-node-log\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473815 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-etc-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473833 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-cni-bin\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473852 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec5c6569-64d4-4591-bbac-b70ebcca836f-proxy-tls\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473876 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cnibin\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473871 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-k8s-cni-cncf-io\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473930 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-cni-multus\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473957 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ec5c6569-64d4-4591-bbac-b70ebcca836f-rootfs\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474015 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474110 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-cnibin\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474154 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c881afa1-c693-450f-912c-cbb4c33bb04f-hosts-file\") pod \"node-resolver-jv7sg\" (UID: \"c881afa1-c693-450f-912c-cbb4c33bb04f\") " pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474223 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-os-release\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474281 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-netns\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474357 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-cni-bin\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474812 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cni-binary-copy\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474874 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474899 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-system-cni-dir\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473851 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-run-multus-certs\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.473899 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-bin\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474935 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-cnibin\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474971 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovn-node-metrics-cert\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.474998 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-systemd\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475018 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475047 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4f75f165-ef56-40be-bfd4-1843de92b356-cni-binary-copy\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475078 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqvhh\" (UniqueName: \"kubernetes.io/projected/c881afa1-c693-450f-912c-cbb4c33bb04f-kube-api-access-nqvhh\") pod \"node-resolver-jv7sg\" (UID: \"c881afa1-c693-450f-912c-cbb4c33bb04f\") " pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475094 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-system-cni-dir\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475122 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-ovn\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475139 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475156 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-kubelet\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475178 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-env-overrides\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475197 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4f75f165-ef56-40be-bfd4-1843de92b356-multus-daemon-config\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475212 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475240 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-netns\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475255 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-slash\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475272 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-etc-kubernetes\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475291 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmffv\" (UniqueName: \"kubernetes.io/projected/4f75f165-ef56-40be-bfd4-1843de92b356-kube-api-access-rmffv\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475310 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec5c6569-64d4-4591-bbac-b70ebcca836f-mcd-auth-proxy-config\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475332 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475351 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-os-release\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475369 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-log-socket\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475466 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475495 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475508 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475520 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475530 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475541 4906 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475553 4906 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475146 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-system-cni-dir\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475584 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475596 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475599 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-etc-kubernetes\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475607 4906 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475626 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4f75f165-ef56-40be-bfd4-1843de92b356-cni-binary-copy\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475640 4906 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475652 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475664 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475702 4906 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475713 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475713 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-os-release\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475725 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475820 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475833 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475844 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475855 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475865 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475875 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475887 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475898 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475907 4906 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475917 4906 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475929 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475539 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f75f165-ef56-40be-bfd4-1843de92b356-host-var-lib-kubelet\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475963 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475980 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.475992 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476004 4906 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476017 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476029 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476043 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476056 4906 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476072 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476128 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476140 4906 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476151 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476163 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476174 4906 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476184 4906 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476197 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476202 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ec5c6569-64d4-4591-bbac-b70ebcca836f-mcd-auth-proxy-config\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476207 4906 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476244 4906 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476254 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476264 4906 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476273 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476282 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476294 4906 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476303 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476315 4906 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476324 4906 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476334 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476344 4906 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476249 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4f75f165-ef56-40be-bfd4-1843de92b356-multus-daemon-config\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476354 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476416 4906 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476436 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476451 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476461 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476471 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476481 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476492 4906 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476505 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476517 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476530 4906 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476544 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476556 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476567 4906 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476576 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476588 4906 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476600 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476613 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476627 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476639 4906 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476651 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476690 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476705 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476717 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476731 4906 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476745 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476758 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476772 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476785 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476798 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476811 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476825 4906 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476840 4906 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476853 4906 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476869 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476882 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476894 4906 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476906 4906 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476919 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476931 4906 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476944 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476957 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476969 4906 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476980 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.476993 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477005 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477018 4906 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477033 4906 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477052 4906 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477064 4906 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477074 4906 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477083 4906 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477094 4906 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477104 4906 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477112 4906 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477121 4906 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477130 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477140 4906 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477151 4906 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477159 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477169 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477180 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477205 4906 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477215 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477224 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477233 4906 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477243 4906 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477253 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477269 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477278 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477291 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477304 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477318 4906 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477329 4906 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477338 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477347 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477356 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477365 4906 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477374 4906 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477384 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477407 4906 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477420 4906 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477431 4906 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477441 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477451 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477461 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477469 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477479 4906 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477490 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477503 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477516 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477529 4906 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477539 4906 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477549 4906 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477558 4906 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477569 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477583 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477597 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477609 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477620 4906 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477631 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477642 4906 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477651 4906 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477662 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477692 4906 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477705 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477717 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477737 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477748 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477759 4906 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477770 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477782 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477794 4906 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477806 4906 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477818 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477830 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477843 4906 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477856 4906 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477869 4906 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477880 4906 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477892 4906 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477903 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477916 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477928 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477939 4906 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477956 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477967 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477979 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.477991 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.478002 4906 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.478012 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.478023 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.478037 4906 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.478047 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.478058 4906 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.478279 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.479286 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46e886ca-acf1-4bf0-94fd-69dd208d8da9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.484216 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec5c6569-64d4-4591-bbac-b70ebcca836f-proxy-tls\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.487072 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.491880 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqvhh\" (UniqueName: \"kubernetes.io/projected/c881afa1-c693-450f-912c-cbb4c33bb04f-kube-api-access-nqvhh\") pod \"node-resolver-jv7sg\" (UID: \"c881afa1-c693-450f-912c-cbb4c33bb04f\") " pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.492335 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4j66\" (UniqueName: \"kubernetes.io/projected/ec5c6569-64d4-4591-bbac-b70ebcca836f-kube-api-access-d4j66\") pod \"machine-config-daemon-2xgvr\" (UID: \"ec5c6569-64d4-4591-bbac-b70ebcca836f\") " pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.492956 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmffv\" (UniqueName: \"kubernetes.io/projected/4f75f165-ef56-40be-bfd4-1843de92b356-kube-api-access-rmffv\") pod \"multus-lm9xt\" (UID: \"4f75f165-ef56-40be-bfd4-1843de92b356\") " pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.497600 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.506370 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7hrg\" (UniqueName: \"kubernetes.io/projected/46e886ca-acf1-4bf0-94fd-69dd208d8da9-kube-api-access-w7hrg\") pod \"multus-additional-cni-plugins-6mz65\" (UID: \"46e886ca-acf1-4bf0-94fd-69dd208d8da9\") " pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.507634 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.515249 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.517592 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8"} Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.523977 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.524448 4906 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.534080 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.544816 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.554644 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.565469 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.576938 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579560 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hvsb\" (UniqueName: \"kubernetes.io/projected/92d6f3b5-c353-4412-8f38-989b53ddef2a-kube-api-access-4hvsb\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579638 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-systemd-units\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579702 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-script-lib\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579739 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579777 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-config\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579814 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-kubelet\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579847 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-var-lib-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579865 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579880 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-node-log\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579915 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-kubelet\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580001 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-etc-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580039 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-bin\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580044 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-var-lib-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580061 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovn-node-metrics-cert\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580005 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-node-log\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580102 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-systemd\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.579857 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-systemd-units\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580121 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580127 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-etc-openvswitch\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580154 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-ovn\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580184 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580195 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-bin\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580209 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-env-overrides\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580237 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580242 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-netns\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580264 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-slash\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580286 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-ovn-kubernetes\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580302 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-log-socket\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580322 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-ovn\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580323 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-netd\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580348 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-netd\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580266 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-systemd\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580420 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-log-socket\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580447 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-netns\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580513 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-slash\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.580568 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-script-lib\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.581023 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-env-overrides\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.581208 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-config\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.583355 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovn-node-metrics-cert\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.587005 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.591233 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lm9xt" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.598149 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hvsb\" (UniqueName: \"kubernetes.io/projected/92d6f3b5-c353-4412-8f38-989b53ddef2a-kube-api-access-4hvsb\") pod \"ovnkube-node-tkkm2\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.599457 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jv7sg" Nov 23 06:50:05 crc kubenswrapper[4906]: W1123 06:50:05.602872 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f75f165_ef56_40be_bfd4_1843de92b356.slice/crio-eba8011f203d967ebd4adbe34dfecfe3ffd25478aa7558ad503fcd528d618221 WatchSource:0}: Error finding container eba8011f203d967ebd4adbe34dfecfe3ffd25478aa7558ad503fcd528d618221: Status 404 returned error can't find the container with id eba8011f203d967ebd4adbe34dfecfe3ffd25478aa7558ad503fcd528d618221 Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.604647 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.605148 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.613951 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.615387 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: W1123 06:50:05.617923 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc881afa1_c693_450f_912c_cbb4c33bb04f.slice/crio-97bb87e8a8f39134c2201bfaf3bdee879f44f938e20bf3ba257aca27cc79a3db WatchSource:0}: Error finding container 97bb87e8a8f39134c2201bfaf3bdee879f44f938e20bf3ba257aca27cc79a3db: Status 404 returned error can't find the container with id 97bb87e8a8f39134c2201bfaf3bdee879f44f938e20bf3ba257aca27cc79a3db Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.619730 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-6mz65" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.628076 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:50:05 crc kubenswrapper[4906]: W1123 06:50:05.629407 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec5c6569_64d4_4591_bbac_b70ebcca836f.slice/crio-f3380bb1d20651aae28b8269718aeb5022c18998164cbc1753ed4baeedc17238 WatchSource:0}: Error finding container f3380bb1d20651aae28b8269718aeb5022c18998164cbc1753ed4baeedc17238: Status 404 returned error can't find the container with id f3380bb1d20651aae28b8269718aeb5022c18998164cbc1753ed4baeedc17238 Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.631255 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.653232 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.704051 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.721357 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.732528 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.734759 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.745026 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.754293 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: W1123 06:50:05.756883 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92d6f3b5_c353_4412_8f38_989b53ddef2a.slice/crio-22628ec75d1275fd31a92082275804ea66d5e70954359338212c45acabe91dfc WatchSource:0}: Error finding container 22628ec75d1275fd31a92082275804ea66d5e70954359338212c45acabe91dfc: Status 404 returned error can't find the container with id 22628ec75d1275fd31a92082275804ea66d5e70954359338212c45acabe91dfc Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.766188 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.775098 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.785335 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.800253 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.811507 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.821824 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.984197 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.984290 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.984319 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984456 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:50:06.984439968 +0000 UTC m=+22.497831271 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.984726 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:05 crc kubenswrapper[4906]: I1123 06:50:05.984750 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984813 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984867 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984920 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:06.984892198 +0000 UTC m=+22.498283501 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984934 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984893 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984949 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984964 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984972 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984982 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:06.984953259 +0000 UTC m=+22.498344562 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.984991 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.985000 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:06.98499293 +0000 UTC m=+22.498384233 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:05 crc kubenswrapper[4906]: E1123 06:50:05.985038 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:06.98503064 +0000 UTC m=+22.498421943 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.522588 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0" exitCode=0 Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.522725 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.522769 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"22628ec75d1275fd31a92082275804ea66d5e70954359338212c45acabe91dfc"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.525710 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.525785 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.525800 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3e6cf41fd83aa3dd9298aea0382716722497626cfc88a0d0db07b1334cce9e66"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.527794 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.527830 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9b16a4e31abc16416d7652dfdd52f4ef81751a7da7537a6fae7039c7a11e2bd4"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.530100 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.530141 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.530161 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"f3380bb1d20651aae28b8269718aeb5022c18998164cbc1753ed4baeedc17238"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.532208 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jv7sg" event={"ID":"c881afa1-c693-450f-912c-cbb4c33bb04f","Type":"ContainerStarted","Data":"af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.532241 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jv7sg" event={"ID":"c881afa1-c693-450f-912c-cbb4c33bb04f","Type":"ContainerStarted","Data":"97bb87e8a8f39134c2201bfaf3bdee879f44f938e20bf3ba257aca27cc79a3db"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.534757 4906 generic.go:334] "Generic (PLEG): container finished" podID="46e886ca-acf1-4bf0-94fd-69dd208d8da9" containerID="901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06" exitCode=0 Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.534826 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerDied","Data":"901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.534852 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerStarted","Data":"f0116865d0a4a2c2971067dff5531fb42f03a77da334b5c5d70431aa50f4bc28"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.537434 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"80f246fa4a24bb1afa1e91be1bebc1ccfef2a71e90a68471db0622265cdc1748"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.539875 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerStarted","Data":"2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.539911 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerStarted","Data":"eba8011f203d967ebd4adbe34dfecfe3ffd25478aa7558ad503fcd528d618221"} Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.540770 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.550899 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.582110 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.598323 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.610573 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.622520 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.635325 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.651700 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.664732 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.682161 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.696105 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.715995 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.730710 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.740818 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.754038 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.764623 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.779896 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.792916 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.809759 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.824485 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.838899 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.863233 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.880766 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.893779 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.905030 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.917624 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.931303 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:06Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.995406 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.995557 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995627 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:50:08.995598655 +0000 UTC m=+24.508989958 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995673 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995712 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995724 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.995719 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.995759 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995772 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:08.995758719 +0000 UTC m=+24.509150022 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:06 crc kubenswrapper[4906]: I1123 06:50:06.995798 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995937 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995973 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:08.995966303 +0000 UTC m=+24.509357606 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.995970 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.996030 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.996093 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:08.996058345 +0000 UTC m=+24.509449658 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.996094 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.996121 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:06 crc kubenswrapper[4906]: E1123 06:50:06.996189 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:08.996167857 +0000 UTC m=+24.509559160 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.192910 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.207791 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.211272 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.211603 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.224960 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.244196 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.285985 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.316489 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.340485 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.355664 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.355760 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.355707 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:07 crc kubenswrapper[4906]: E1123 06:50:07.355838 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:07 crc kubenswrapper[4906]: E1123 06:50:07.355999 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:07 crc kubenswrapper[4906]: E1123 06:50:07.356254 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.358781 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.360211 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.361097 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.362334 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.363043 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.364116 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.364727 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.365375 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.366488 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.367180 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.368138 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.368708 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.370126 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.370954 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.371501 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.372120 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.372746 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.373371 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.373707 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.374907 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.375502 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.376499 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.377091 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.377901 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.378965 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.379857 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.380830 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.381704 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.382795 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.383311 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.384419 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.384978 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.385578 4906 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.385746 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.386750 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.387877 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.388458 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.389421 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.390952 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.391602 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.392554 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.393222 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.394322 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.394852 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.396074 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.396767 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.397767 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.398282 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.399533 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.399564 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.400139 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.401535 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.402086 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.403026 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.403512 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.404099 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.405080 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.405587 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.416468 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.430999 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.441161 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.454115 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.470492 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.483706 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.507413 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.537034 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.546410 4906 generic.go:334] "Generic (PLEG): container finished" podID="46e886ca-acf1-4bf0-94fd-69dd208d8da9" containerID="468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f" exitCode=0 Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.546496 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerDied","Data":"468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f"} Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.563300 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.563532 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.563616 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.563637 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.563655 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.563675 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.599737 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-pd47s"] Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.600837 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.609700 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.616432 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.634896 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.655907 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.673813 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.702549 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-serviceca\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.702637 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snpzh\" (UniqueName: \"kubernetes.io/projected/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-kube-api-access-snpzh\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.702720 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-host\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.745154 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.771623 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.803130 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-serviceca\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.803190 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snpzh\" (UniqueName: \"kubernetes.io/projected/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-kube-api-access-snpzh\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.803239 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-host\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.803306 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-host\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.804298 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-serviceca\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.806036 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.833317 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snpzh\" (UniqueName: \"kubernetes.io/projected/407b04ac-8bbb-4112-88bc-bbfc89e37f6d-kube-api-access-snpzh\") pod \"node-ca-pd47s\" (UID: \"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\") " pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.865231 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.906970 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.919653 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pd47s" Nov 23 06:50:07 crc kubenswrapper[4906]: W1123 06:50:07.930867 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod407b04ac_8bbb_4112_88bc_bbfc89e37f6d.slice/crio-65781301deda88a8c62da45294c0f452f0b816a39600dd39312987b95b4b6195 WatchSource:0}: Error finding container 65781301deda88a8c62da45294c0f452f0b816a39600dd39312987b95b4b6195: Status 404 returned error can't find the container with id 65781301deda88a8c62da45294c0f452f0b816a39600dd39312987b95b4b6195 Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.941236 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:07 crc kubenswrapper[4906]: I1123 06:50:07.982531 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.024857 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.063986 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.101883 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.142777 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.185485 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.220523 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.265201 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.302004 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.349104 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.382093 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.427513 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.467584 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.502840 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.558418 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.574949 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756"} Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.577127 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pd47s" event={"ID":"407b04ac-8bbb-4112-88bc-bbfc89e37f6d","Type":"ContainerStarted","Data":"26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d"} Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.577175 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pd47s" event={"ID":"407b04ac-8bbb-4112-88bc-bbfc89e37f6d","Type":"ContainerStarted","Data":"65781301deda88a8c62da45294c0f452f0b816a39600dd39312987b95b4b6195"} Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.580048 4906 generic.go:334] "Generic (PLEG): container finished" podID="46e886ca-acf1-4bf0-94fd-69dd208d8da9" containerID="30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365" exitCode=0 Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.580212 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerDied","Data":"30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365"} Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.586297 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.611245 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.632222 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.666782 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.709030 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.763334 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.789025 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.829358 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.884328 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.926407 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.948088 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:08 crc kubenswrapper[4906]: I1123 06:50:08.991177 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.018992 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.019153 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019171 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:50:13.019146535 +0000 UTC m=+28.532537828 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.019223 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.019295 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.019342 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019354 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019384 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019408 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019465 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019488 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019490 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:13.019464361 +0000 UTC m=+28.532855694 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019523 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019550 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019484 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019595 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:13.019586113 +0000 UTC m=+28.532977416 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019639 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:13.019627074 +0000 UTC m=+28.533018607 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.019655 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:13.019645614 +0000 UTC m=+28.533037157 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.025063 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.063301 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.105633 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.149797 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.184746 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.356375 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.356407 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.356384 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.356507 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.356597 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:09 crc kubenswrapper[4906]: E1123 06:50:09.356640 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.601495 4906 generic.go:334] "Generic (PLEG): container finished" podID="46e886ca-acf1-4bf0-94fd-69dd208d8da9" containerID="9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1" exitCode=0 Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.601869 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerDied","Data":"9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1"} Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.619429 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.638532 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.659306 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.677893 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.696899 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.716275 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.735651 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.753165 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.772827 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.798014 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.815729 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.835982 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.851782 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.867156 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:09 crc kubenswrapper[4906]: I1123 06:50:09.883170 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.613225 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.615847 4906 generic.go:334] "Generic (PLEG): container finished" podID="46e886ca-acf1-4bf0-94fd-69dd208d8da9" containerID="8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60" exitCode=0 Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.615915 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerDied","Data":"8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.636557 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.667058 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.688868 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.705125 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.719744 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.738116 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.750696 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.752527 4906 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.754653 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.754716 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.754728 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.754913 4906 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.763258 4906 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.763533 4906 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.771594 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.771838 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.771907 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.771991 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.772053 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:10Z","lastTransitionTime":"2025-11-23T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.785140 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: E1123 06:50:10.785340 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.788875 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.788969 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.789026 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.789086 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.789153 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:10Z","lastTransitionTime":"2025-11-23T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.799142 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: E1123 06:50:10.801822 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.806184 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.806283 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.806310 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.806347 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.806374 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:10Z","lastTransitionTime":"2025-11-23T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.812356 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: E1123 06:50:10.819106 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.823524 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.823582 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.823596 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.823621 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.823638 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:10Z","lastTransitionTime":"2025-11-23T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.829119 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: E1123 06:50:10.835753 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.839436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.839481 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.839495 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.839516 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.839534 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:10Z","lastTransitionTime":"2025-11-23T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.845036 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: E1123 06:50:10.851536 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: E1123 06:50:10.851643 4906 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.853589 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.853638 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.853650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.853695 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.853708 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:10Z","lastTransitionTime":"2025-11-23T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.857640 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.869610 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.881105 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:10Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.959232 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.959288 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.959304 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.959325 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:10 crc kubenswrapper[4906]: I1123 06:50:10.959337 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:10Z","lastTransitionTime":"2025-11-23T06:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.062356 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.062408 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.062418 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.062449 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.062460 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.165553 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.165626 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.165650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.165712 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.165733 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.268356 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.268436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.268458 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.268491 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.268517 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.356721 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.356762 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.356863 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:11 crc kubenswrapper[4906]: E1123 06:50:11.356931 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:11 crc kubenswrapper[4906]: E1123 06:50:11.357066 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:11 crc kubenswrapper[4906]: E1123 06:50:11.357359 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.375785 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.375877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.375896 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.375923 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.375942 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.479166 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.479202 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.479211 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.479226 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.479235 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.583751 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.583800 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.583812 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.583832 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.583847 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.634617 4906 generic.go:334] "Generic (PLEG): container finished" podID="46e886ca-acf1-4bf0-94fd-69dd208d8da9" containerID="815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71" exitCode=0 Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.634744 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerDied","Data":"815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.660276 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.680504 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.688156 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.688230 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.688246 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.688267 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.688315 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.698697 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.720555 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.734984 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.776470 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.791754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.791805 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.791817 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.791836 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.791847 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.802803 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.821652 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.841108 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.857613 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.872782 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.890345 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.895079 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.895127 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.895140 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.895161 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.895180 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.902905 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.920178 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.934395 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:11Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.998507 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.998556 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.998570 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.998593 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:11 crc kubenswrapper[4906]: I1123 06:50:11.998610 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:11Z","lastTransitionTime":"2025-11-23T06:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.101269 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.101310 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.101322 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.101340 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.101356 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.204485 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.204527 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.204539 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.204563 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.204575 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.308221 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.308270 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.308284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.308306 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.308321 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.411268 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.411333 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.411353 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.411379 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.411396 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.513810 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.513855 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.513864 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.513881 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.513892 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.616865 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.616904 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.616912 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.616946 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.616958 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.642784 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" event={"ID":"46e886ca-acf1-4bf0-94fd-69dd208d8da9","Type":"ContainerStarted","Data":"b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.648834 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.653043 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.653126 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.653152 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.663232 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.691770 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.696045 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.696142 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.719663 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.720330 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.720370 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.720386 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.720409 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.720428 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.736421 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.754050 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.770708 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.791289 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.812429 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.823338 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.823411 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.823425 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.823472 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.823489 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.827527 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.843164 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.890861 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.910554 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.922345 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.925863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.925920 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.925932 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.925953 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.925967 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:12Z","lastTransitionTime":"2025-11-23T06:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.942649 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.956206 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.973152 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:12 crc kubenswrapper[4906]: I1123 06:50:12.987411 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.000605 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.021790 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.029545 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.029596 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.029610 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.029628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.029645 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.037917 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.051358 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.065906 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.066065 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.066095 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066145 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:50:21.06612168 +0000 UTC m=+36.579512983 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.066183 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.066216 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066263 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066274 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066286 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066293 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066299 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066303 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066305 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066339 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:21.066328874 +0000 UTC m=+36.579720177 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066356 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:21.066348715 +0000 UTC m=+36.579740028 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066355 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066374 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:21.066362106 +0000 UTC m=+36.579753409 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.066461 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:21.066441578 +0000 UTC m=+36.579832881 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.067640 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.081490 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.093577 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.105370 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.121215 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.131921 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.131970 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.131982 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.132001 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.132014 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.135022 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.153964 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.171088 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.182138 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.241723 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.241775 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.241787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.241813 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.241826 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.344260 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.344305 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.344314 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.344327 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.344335 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.355524 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.355524 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.355620 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.355714 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.355901 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:13 crc kubenswrapper[4906]: E1123 06:50:13.356542 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.446928 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.446986 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.446995 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.447009 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.447231 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.550396 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.550444 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.550459 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.550480 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.550497 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.653577 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.653645 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.653661 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.653715 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.653734 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.757213 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.757253 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.757265 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.757284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.757296 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.860014 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.860067 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.860081 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.860102 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.860118 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.963225 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.963275 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.963285 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.963300 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:13 crc kubenswrapper[4906]: I1123 06:50:13.963312 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:13Z","lastTransitionTime":"2025-11-23T06:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.066485 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.066544 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.066559 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.066585 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.066602 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.169149 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.169241 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.169263 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.169288 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.169317 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.271738 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.271768 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.271776 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.271788 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.271797 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.374647 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.374735 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.374753 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.374776 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.374794 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.476978 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.477008 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.477019 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.477038 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.477050 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.579335 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.579366 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.579374 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.579386 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.579395 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.681313 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.681350 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.681358 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.681373 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.681387 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.783421 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.783455 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.783466 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.783481 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.783490 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.886096 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.886137 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.886147 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.886163 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.886174 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.989066 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.989156 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.989173 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.989199 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:14 crc kubenswrapper[4906]: I1123 06:50:14.989217 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:14Z","lastTransitionTime":"2025-11-23T06:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.091232 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.091524 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.091538 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.091553 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.091564 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.193743 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.193806 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.193819 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.193834 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.193845 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.296593 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.296634 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.296650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.296668 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.296708 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.356365 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.356396 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.356365 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:15 crc kubenswrapper[4906]: E1123 06:50:15.356495 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:15 crc kubenswrapper[4906]: E1123 06:50:15.356565 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:15 crc kubenswrapper[4906]: E1123 06:50:15.356632 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.375452 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.398756 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.398843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.398862 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.398893 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.398914 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.402841 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.434575 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.451867 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.468250 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.486180 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.499063 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.500713 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.500828 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.500836 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.500850 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.500859 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.511620 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.530303 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.547041 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.563127 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.581482 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.597060 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.603844 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.603876 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.603901 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.603918 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.603927 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.610531 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.621953 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.658904 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/0.log" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.661518 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554" exitCode=1 Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.661565 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.662412 4906 scope.go:117] "RemoveContainer" containerID="9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.675161 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.705624 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.705665 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.705697 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.705715 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.705727 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.707895 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.721525 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.732642 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.754037 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.770213 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.783378 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.807973 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.808011 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.808022 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.808039 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.808052 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.814659 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:15Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1123 06:50:14.992598 6226 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:14.992619 6226 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:14.992647 6226 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:14.992714 6226 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:14.993353 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:14.993377 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:14.993382 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:14.993423 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:14.993460 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:14.993489 6226 factory.go:656] Stopping watch factory\\\\nI1123 06:50:14.993499 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:14.993506 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:14.993552 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:14.993568 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:14.993629 6226 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.838539 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.854040 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.867598 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.901493 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.910127 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.910156 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.910164 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.910179 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.910189 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:15Z","lastTransitionTime":"2025-11-23T06:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.915309 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.925766 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:15 crc kubenswrapper[4906]: I1123 06:50:15.936055 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.013502 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.013551 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.013566 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.013590 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.013605 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.116056 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.116107 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.116122 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.116144 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.116161 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.218972 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.219237 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.219247 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.219264 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.219275 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.321507 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.321549 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.321559 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.321577 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.321587 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.423493 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.423542 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.423551 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.423567 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.423579 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.527427 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.527461 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.527469 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.527484 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.527493 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.630225 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.630259 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.630271 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.630287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.630297 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.667160 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/0.log" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.669959 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.670363 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.684384 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.706950 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.727228 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.732672 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.732764 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.732779 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.732804 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.732817 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.749375 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.768901 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.796337 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:15Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1123 06:50:14.992598 6226 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:14.992619 6226 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:14.992647 6226 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:14.992714 6226 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:14.993353 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:14.993377 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:14.993382 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:14.993423 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:14.993460 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:14.993489 6226 factory.go:656] Stopping watch factory\\\\nI1123 06:50:14.993499 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:14.993506 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:14.993552 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:14.993568 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:14.993629 6226 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.829091 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.836128 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.836168 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.836179 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.836201 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.836213 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.847105 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.865341 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.883428 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.911521 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.931307 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.939404 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.939436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.939447 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.939462 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.939471 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:16Z","lastTransitionTime":"2025-11-23T06:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.950407 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.973661 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:16 crc kubenswrapper[4906]: I1123 06:50:16.989980 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:16Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.043779 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.043840 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.043854 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.043875 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.043892 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.147225 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.147260 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.147271 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.147288 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.147298 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.249593 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.249628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.249637 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.249650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.249661 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.351590 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.351669 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.351724 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.351752 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.351770 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.355883 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.355942 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.356027 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:17 crc kubenswrapper[4906]: E1123 06:50:17.356071 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:17 crc kubenswrapper[4906]: E1123 06:50:17.356299 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:17 crc kubenswrapper[4906]: E1123 06:50:17.356342 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.455257 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.455286 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.455293 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.455307 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.455316 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.558793 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.558840 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.558856 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.558881 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.558897 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.665732 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.665769 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.665778 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.665793 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.665803 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.674558 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/1.log" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.675427 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/0.log" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.678945 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a" exitCode=1 Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.678996 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.679096 4906 scope.go:117] "RemoveContainer" containerID="9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.679844 4906 scope.go:117] "RemoveContainer" containerID="7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a" Nov 23 06:50:17 crc kubenswrapper[4906]: E1123 06:50:17.680028 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.695064 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.708639 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.723844 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.738131 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.753281 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.768838 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.768887 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.768900 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.768920 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.768930 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.769317 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.786734 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:15Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1123 06:50:14.992598 6226 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:14.992619 6226 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:14.992647 6226 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:14.992714 6226 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:14.993353 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:14.993377 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:14.993382 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:14.993423 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:14.993460 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:14.993489 6226 factory.go:656] Stopping watch factory\\\\nI1123 06:50:14.993499 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:14.993506 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:14.993552 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:14.993568 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:14.993629 6226 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.811335 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.831873 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.842165 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk"] Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.842872 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.845273 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.845515 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.856598 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.871286 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.871445 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.871566 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.871674 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.871898 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.906364 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.928808 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjkzg\" (UniqueName: \"kubernetes.io/projected/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-kube-api-access-bjkzg\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.929046 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.929251 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.929381 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.931551 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.948941 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.964556 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.974774 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.974837 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.975046 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.975074 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.975089 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:17Z","lastTransitionTime":"2025-11-23T06:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:17 crc kubenswrapper[4906]: I1123 06:50:17.981159 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.014786 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.030743 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjkzg\" (UniqueName: \"kubernetes.io/projected/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-kube-api-access-bjkzg\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.030812 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.030869 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.030898 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.031592 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.032012 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.037171 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.038516 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.055265 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjkzg\" (UniqueName: \"kubernetes.io/projected/b8b0a056-21ed-4d43-bde9-5ecff043aaa4-kube-api-access-bjkzg\") pod \"ovnkube-control-plane-749d76644c-6tlzk\" (UID: \"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.058386 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.078240 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.078157 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.078301 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.078316 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.078339 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.078716 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.094796 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.108858 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.120747 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.133901 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.147582 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.162739 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.174550 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b265751acefc6fd597d0d3dc77b15d3edb9dc55d132c63884d273ac6c5d0554\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:15Z\\\",\\\"message\\\":\\\"ressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1123 06:50:14.992598 6226 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:14.992619 6226 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:14.992647 6226 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:14.992714 6226 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:14.993353 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:14.993377 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:14.993382 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:14.993423 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:14.993460 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:14.993489 6226 factory.go:656] Stopping watch factory\\\\nI1123 06:50:14.993499 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:14.993506 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:14.993552 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:14.993568 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:14.993629 6226 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: W1123 06:50:18.182023 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8b0a056_21ed_4d43_bde9_5ecff043aaa4.slice/crio-c5f27c14b0d8c054d21b646f89a86f4a9a5efb638244e5383aa653d4a8ebec20 WatchSource:0}: Error finding container c5f27c14b0d8c054d21b646f89a86f4a9a5efb638244e5383aa653d4a8ebec20: Status 404 returned error can't find the container with id c5f27c14b0d8c054d21b646f89a86f4a9a5efb638244e5383aa653d4a8ebec20 Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.182846 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.182887 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.182898 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.182921 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.182937 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.197047 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.216652 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.235313 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.253306 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.269825 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.286063 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.286618 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.286647 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.286660 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.286693 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.286703 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.390887 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.390938 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.390956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.390985 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.391011 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.493883 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.493920 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.493928 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.493941 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.493950 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.596731 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.596776 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.596788 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.596804 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.596815 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.683971 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/1.log" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.689093 4906 scope.go:117] "RemoveContainer" containerID="7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a" Nov 23 06:50:18 crc kubenswrapper[4906]: E1123 06:50:18.689400 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.690568 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" event={"ID":"b8b0a056-21ed-4d43-bde9-5ecff043aaa4","Type":"ContainerStarted","Data":"faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.690639 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" event={"ID":"b8b0a056-21ed-4d43-bde9-5ecff043aaa4","Type":"ContainerStarted","Data":"606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.690657 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" event={"ID":"b8b0a056-21ed-4d43-bde9-5ecff043aaa4","Type":"ContainerStarted","Data":"c5f27c14b0d8c054d21b646f89a86f4a9a5efb638244e5383aa653d4a8ebec20"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.698928 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.699000 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.699023 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.699048 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.699064 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.702548 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.715845 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.726074 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.754996 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.768529 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.783408 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.803877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.803930 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.803942 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.803965 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.803979 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.807270 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.822736 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.836674 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.857145 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.870328 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.885620 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.898998 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.907398 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.907456 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.907471 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.907501 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.907516 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:18Z","lastTransitionTime":"2025-11-23T06:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.917532 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.931853 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.944839 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.958834 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.972591 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:18 crc kubenswrapper[4906]: I1123 06:50:18.985558 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.000270 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-8drn8"] Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.001007 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.001103 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.009641 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.009713 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.009728 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.009747 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.009762 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.009595 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.030406 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.042383 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.045290 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.045387 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn5t8\" (UniqueName: \"kubernetes.io/projected/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-kube-api-access-dn5t8\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.056754 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.069696 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.087048 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.099029 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.112869 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.112948 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.112976 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.113014 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.113042 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.113992 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.131867 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.146605 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.146709 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn5t8\" (UniqueName: \"kubernetes.io/projected/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-kube-api-access-dn5t8\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.146783 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.146878 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:19.646853957 +0000 UTC m=+35.160245260 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.152735 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.173022 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn5t8\" (UniqueName: \"kubernetes.io/projected/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-kube-api-access-dn5t8\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.180067 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.196542 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.205948 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.215047 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.215120 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.215134 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.215152 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.215453 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.221562 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.234627 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.249094 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.274095 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.294431 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.313111 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.317988 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.318055 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.318068 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.318088 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.318101 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.327247 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.338346 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.352582 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.355672 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.355813 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.355824 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.355935 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.356030 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.356189 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.369543 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.385440 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.405852 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.421772 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.421863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.421880 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.421906 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.421924 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.425548 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.441988 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.459299 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.478464 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.497137 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.524892 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.524956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.524969 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.524994 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.525008 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.628072 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.628151 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.628170 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.628214 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.628232 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.652049 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.652337 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:19 crc kubenswrapper[4906]: E1123 06:50:19.652487 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:20.652448885 +0000 UTC m=+36.165840228 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.730977 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.731026 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.731037 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.731054 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.731064 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.834287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.834387 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.834415 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.834458 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.834489 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.940468 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.940542 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.940954 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.941076 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.941120 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.941142 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:19Z","lastTransitionTime":"2025-11-23T06:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.958219 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.974987 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:19 crc kubenswrapper[4906]: I1123 06:50:19.989239 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.022097 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.038144 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.045746 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.045803 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.045819 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.045841 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.045858 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.053639 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.073789 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.092342 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.108371 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.136302 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.148836 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.148893 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.148910 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.148935 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.148951 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.152364 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.173206 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.192729 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.210758 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.229175 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.242475 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.252105 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.252198 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.252270 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.252308 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.252331 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.258145 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.355733 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.356119 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.356390 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.356471 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.356556 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.356636 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: E1123 06:50:20.357110 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.461101 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.461175 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.461191 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.461215 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.461230 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.565375 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.565430 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.565440 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.565462 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.565473 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.664544 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:20 crc kubenswrapper[4906]: E1123 06:50:20.664863 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:20 crc kubenswrapper[4906]: E1123 06:50:20.664996 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:22.664967441 +0000 UTC m=+38.178358744 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.668089 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.668151 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.668171 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.668198 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.668215 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.771213 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.771284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.771302 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.771341 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.771361 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.876057 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.876123 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.876137 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.876163 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.876187 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.980255 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.980355 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.980384 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.980419 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:20 crc kubenswrapper[4906]: I1123 06:50:20.980443 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:20Z","lastTransitionTime":"2025-11-23T06:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.070267 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.070481 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.070739 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.070839 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.070915 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071127 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071179 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:50:37.071122329 +0000 UTC m=+52.584513672 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071196 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071267 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071306 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:37.071284502 +0000 UTC m=+52.584675845 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071350 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071366 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:37.071330523 +0000 UTC m=+52.584721856 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071195 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071392 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071410 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071428 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071454 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:37.071441425 +0000 UTC m=+52.584832768 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.071540 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:37.071504627 +0000 UTC m=+52.584896010 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.084754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.084811 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.084830 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.084863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.084882 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.162284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.162357 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.162375 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.162405 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.162426 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.188190 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.195379 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.195464 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.195487 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.195522 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.195545 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.227924 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.234260 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.234337 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.234361 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.234394 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.234418 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.257231 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.264019 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.264067 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.264089 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.264117 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.264138 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.286927 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.293288 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.293399 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.293417 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.293440 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.293456 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.316625 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.316791 4906 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.319633 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.319668 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.319697 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.319716 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.319727 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.356451 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.356589 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.356667 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.356835 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.357070 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:21 crc kubenswrapper[4906]: E1123 06:50:21.357201 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.422906 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.422985 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.423003 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.423030 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.423049 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.527098 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.527202 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.527235 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.527275 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.527302 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.630558 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.630631 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.630650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.630719 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.630741 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.734840 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.734923 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.734947 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.734979 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.735000 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.838661 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.838754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.838773 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.838806 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.838826 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.941874 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.941976 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.942001 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.942038 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:21 crc kubenswrapper[4906]: I1123 06:50:21.942063 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:21Z","lastTransitionTime":"2025-11-23T06:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.046419 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.046479 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.046497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.046525 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.046545 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.149959 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.150039 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.150057 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.150095 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.150120 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.254537 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.254639 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.254668 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.254749 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.254782 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.356541 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:22 crc kubenswrapper[4906]: E1123 06:50:22.356968 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.358660 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.358783 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.358823 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.358857 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.358883 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.464108 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.464218 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.464239 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.464272 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.464295 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.568043 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.568132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.568157 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.568189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.568213 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.676862 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.677427 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.677461 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.677495 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.677517 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.699761 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:22 crc kubenswrapper[4906]: E1123 06:50:22.699941 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:22 crc kubenswrapper[4906]: E1123 06:50:22.700051 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:26.700028178 +0000 UTC m=+42.213419491 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.784929 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.785010 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.785031 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.785061 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.785080 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.889169 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.889294 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.889317 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.889347 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.889368 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.991817 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.991866 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.991877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.991896 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:22 crc kubenswrapper[4906]: I1123 06:50:22.991913 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:22Z","lastTransitionTime":"2025-11-23T06:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.096391 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.096462 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.096486 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.096516 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.096538 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.201309 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.201419 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.201443 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.201479 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.201507 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.306239 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.306329 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.306355 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.306389 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.306415 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.356169 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.356243 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.356303 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:23 crc kubenswrapper[4906]: E1123 06:50:23.356426 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:23 crc kubenswrapper[4906]: E1123 06:50:23.356779 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:23 crc kubenswrapper[4906]: E1123 06:50:23.356896 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.410421 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.410491 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.410512 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.410540 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.410561 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.514303 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.514358 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.514376 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.514402 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.514422 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.618749 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.618806 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.618822 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.618845 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.618858 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.722122 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.722182 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.722199 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.722226 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.722245 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.825873 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.825952 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.825970 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.826001 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.826022 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.929569 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.929626 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.929647 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.929673 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:23 crc kubenswrapper[4906]: I1123 06:50:23.929712 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:23Z","lastTransitionTime":"2025-11-23T06:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.032560 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.032600 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.032609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.032625 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.032635 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.135533 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.135588 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.135606 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.135633 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.135651 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.239987 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.240063 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.240081 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.240120 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.240136 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.344745 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.344838 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.344862 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.344894 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.344915 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.355979 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:24 crc kubenswrapper[4906]: E1123 06:50:24.356273 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.449550 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.449644 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.449668 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.449745 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.449768 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.553456 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.553527 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.553547 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.553577 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.553600 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.657845 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.657926 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.657944 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.657973 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.657990 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.760895 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.760982 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.761009 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.761044 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.761073 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.865287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.865385 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.865408 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.865445 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.865469 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.968797 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.968865 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.968883 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.968912 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:24 crc kubenswrapper[4906]: I1123 06:50:24.968933 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:24Z","lastTransitionTime":"2025-11-23T06:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.073070 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.073569 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.073761 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.073934 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.074076 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.178663 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.179074 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.179321 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.179529 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.179750 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.282803 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.282865 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.282884 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.282916 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.282938 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.356241 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.356296 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:25 crc kubenswrapper[4906]: E1123 06:50:25.356455 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:25 crc kubenswrapper[4906]: E1123 06:50:25.356611 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.356811 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:25 crc kubenswrapper[4906]: E1123 06:50:25.356997 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.388889 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.389645 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.389838 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.389876 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.389895 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.391847 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.407170 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.432899 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.455920 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.479504 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.492341 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.492423 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.492450 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.492485 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.492512 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.507547 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.525358 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.542335 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.558531 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.579880 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.596149 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.596234 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.596262 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.596370 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.596425 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.603628 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.624961 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.661227 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.700336 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.700396 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.700414 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.700446 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.700466 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.704945 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.727085 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.746648 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.766957 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.804267 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.804350 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.804370 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.804403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.804423 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.908332 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.908402 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.908424 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.908451 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:25 crc kubenswrapper[4906]: I1123 06:50:25.908473 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:25Z","lastTransitionTime":"2025-11-23T06:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.012151 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.012238 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.012263 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.012298 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.012322 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.116071 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.116139 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.116158 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.116185 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.116202 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.219922 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.219999 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.220013 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.220033 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.220047 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.324164 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.324251 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.324271 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.324314 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.324334 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.355505 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:26 crc kubenswrapper[4906]: E1123 06:50:26.356164 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.428402 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.428996 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.429233 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.429407 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.429559 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.534831 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.535363 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.535502 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.535649 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.535827 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.640390 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.640457 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.640474 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.640499 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.640516 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.743566 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.744045 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.744809 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.744927 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.744959 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.755189 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:26 crc kubenswrapper[4906]: E1123 06:50:26.755421 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:26 crc kubenswrapper[4906]: E1123 06:50:26.755545 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:34.75551031 +0000 UTC m=+50.268901833 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.849206 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.849271 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.849287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.849314 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.849331 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.953336 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.953397 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.953410 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.953436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:26 crc kubenswrapper[4906]: I1123 06:50:26.953452 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:26Z","lastTransitionTime":"2025-11-23T06:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.057497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.057575 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.057594 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.057633 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.057673 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.161507 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.161599 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.161619 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.161652 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.161725 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.265195 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.265271 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.265292 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.265322 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.265341 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.356314 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.356357 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.356529 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:27 crc kubenswrapper[4906]: E1123 06:50:27.356630 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:27 crc kubenswrapper[4906]: E1123 06:50:27.356925 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:27 crc kubenswrapper[4906]: E1123 06:50:27.357052 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.370034 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.370092 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.370104 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.370124 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.370138 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.473650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.474207 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.474336 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.474473 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.474607 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.578031 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.578143 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.578162 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.578196 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.578217 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.681793 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.681848 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.681863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.681882 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.681892 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.785354 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.785435 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.785456 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.785490 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.785509 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.888763 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.888810 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.888820 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.888863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.888875 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.993057 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.993129 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.993145 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.993178 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:27 crc kubenswrapper[4906]: I1123 06:50:27.993196 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:27Z","lastTransitionTime":"2025-11-23T06:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.096532 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.096607 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.096625 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.096652 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.096669 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.200544 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.200634 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.200649 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.200674 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.200715 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.304874 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.304992 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.305022 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.305060 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.305082 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.355499 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:28 crc kubenswrapper[4906]: E1123 06:50:28.356000 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.408308 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.408375 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.408390 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.408415 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.408430 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.513123 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.513215 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.513238 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.513270 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.513289 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.617295 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.617423 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.617454 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.617499 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.617525 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.722544 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.722646 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.722666 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.722727 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.722761 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.826812 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.826891 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.826913 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.826945 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.826965 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.930808 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.930911 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.930937 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.930978 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:28 crc kubenswrapper[4906]: I1123 06:50:28.931000 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:28Z","lastTransitionTime":"2025-11-23T06:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.036509 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.036589 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.036607 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.036637 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.036657 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.139923 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.139998 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.140013 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.140038 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.140051 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.245568 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.245632 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.245653 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.245716 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.245735 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.349612 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.349734 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.349755 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.349795 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.349816 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.356304 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.356384 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:29 crc kubenswrapper[4906]: E1123 06:50:29.356614 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:29 crc kubenswrapper[4906]: E1123 06:50:29.356781 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.358877 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.359087 4906 scope.go:117] "RemoveContainer" containerID="7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a" Nov 23 06:50:29 crc kubenswrapper[4906]: E1123 06:50:29.359086 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.454279 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.454340 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.454355 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.454381 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.454395 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.558418 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.558492 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.558515 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.558550 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.558573 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.665129 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.665182 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.665202 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.665235 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.665259 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.747763 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/1.log" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.753891 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.754530 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.770000 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.770072 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.770092 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.770123 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.770146 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.778316 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.804203 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.827931 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.854160 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.875058 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.875190 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.875212 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.875280 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.875300 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.882555 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.908122 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.928002 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.952175 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.975955 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.978295 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.978364 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.978376 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.978403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:29 crc kubenswrapper[4906]: I1123 06:50:29.978412 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:29Z","lastTransitionTime":"2025-11-23T06:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.002502 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.018348 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.034280 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.049464 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.064225 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.081221 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.081285 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.081299 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.081324 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.081339 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.088482 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.115721 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.131980 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.184966 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.185020 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.185050 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.185071 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.185085 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.288810 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.288878 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.288891 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.288912 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.288925 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.355941 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:30 crc kubenswrapper[4906]: E1123 06:50:30.356845 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.392580 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.392660 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.392709 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.392741 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.392785 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.496479 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.496555 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.496573 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.496607 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.496630 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.600722 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.600798 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.600807 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.600826 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.600836 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.704809 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.704868 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.704880 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.704899 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.705244 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.761923 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/2.log" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.763390 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/1.log" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.768153 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3" exitCode=1 Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.768224 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.768278 4906 scope.go:117] "RemoveContainer" containerID="7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.775210 4906 scope.go:117] "RemoveContainer" containerID="a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3" Nov 23 06:50:30 crc kubenswrapper[4906]: E1123 06:50:30.775845 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.782553 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.804133 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.809843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.809910 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.809929 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.809961 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.809985 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.818953 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.852083 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a41c28bb401226fdf603fd8409cbf5260c377a65c1dcd63fe886d35cf9eb98a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:16Z\\\",\\\"message\\\":\\\"87cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1123 06:50:16.619734 6347 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.872920 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.886879 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.899522 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.913134 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.913188 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.913204 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.913228 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.913244 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:30Z","lastTransitionTime":"2025-11-23T06:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.915154 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.930369 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.949071 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.973706 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:30 crc kubenswrapper[4906]: I1123 06:50:30.991751 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.012371 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.017042 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.017115 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.017128 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.017152 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.017168 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.032424 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.059531 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.074948 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.088610 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.119955 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.120029 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.120044 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.120068 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.120103 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.222781 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.222860 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.222884 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.222922 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.222946 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.327166 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.327273 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.327305 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.327343 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.327369 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.355768 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.355819 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.355778 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.355968 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.356158 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.356389 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.430465 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.430533 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.430553 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.430580 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.430603 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.499326 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.499423 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.499451 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.499487 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.499511 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.517207 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.521982 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.522040 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.522059 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.522088 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.522111 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.541668 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.546772 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.546804 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.546821 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.546843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.546860 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.566765 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.572893 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.572950 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.572964 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.572990 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.573007 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.593069 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.599729 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.599809 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.599836 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.599873 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.599897 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.622518 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.622648 4906 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.625190 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.625245 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.625264 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.625287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.625302 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.728332 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.728403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.728421 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.728449 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.728468 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.776134 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/2.log" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.781482 4906 scope.go:117] "RemoveContainer" containerID="a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3" Nov 23 06:50:31 crc kubenswrapper[4906]: E1123 06:50:31.781816 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.801734 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.825073 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.831887 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.831938 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.831957 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.831981 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.831998 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.846174 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.864369 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.883015 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.899329 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.932228 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.935642 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.935794 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.935818 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.935850 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.935877 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:31Z","lastTransitionTime":"2025-11-23T06:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.957094 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.975605 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:31 crc kubenswrapper[4906]: I1123 06:50:31.993126 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.019379 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.039661 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.039759 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.039780 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.039809 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.039829 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.042906 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.063368 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.084968 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.107352 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.123081 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.140275 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.142950 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.143033 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.143061 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.143100 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.143128 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.246050 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.246129 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.246149 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.246182 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.246207 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.349540 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.349589 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.349599 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.349617 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.349627 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.356031 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:32 crc kubenswrapper[4906]: E1123 06:50:32.356196 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.452639 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.452764 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.452785 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.452818 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.452838 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.554708 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.554737 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.554745 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.554758 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.554768 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.657866 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.657936 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.657956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.657987 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.658007 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.761596 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.761720 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.761740 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.761770 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.761789 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.866222 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.866298 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.866320 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.866384 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.866410 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.970735 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.970826 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.970853 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.970885 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:32 crc kubenswrapper[4906]: I1123 06:50:32.970906 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:32Z","lastTransitionTime":"2025-11-23T06:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.075189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.075257 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.075277 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.075306 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.075327 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.178450 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.178519 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.178538 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.178571 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.178590 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.282620 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.282739 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.282777 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.282815 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.282842 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.355777 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:33 crc kubenswrapper[4906]: E1123 06:50:33.356012 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.356462 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.356567 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:33 crc kubenswrapper[4906]: E1123 06:50:33.357128 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:33 crc kubenswrapper[4906]: E1123 06:50:33.357313 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.385509 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.385599 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.385620 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.385648 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.385669 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.489370 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.489598 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.489660 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.489755 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.489820 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.593754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.594040 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.594101 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.594163 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.594219 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.698298 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.698404 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.698424 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.698454 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.698527 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.801996 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.802066 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.802081 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.802110 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.802125 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.906041 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.906102 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.906114 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.906154 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:33 crc kubenswrapper[4906]: I1123 06:50:33.906169 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:33Z","lastTransitionTime":"2025-11-23T06:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.009621 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.009666 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.009690 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.009706 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.009715 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.112150 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.112234 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.112247 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.112265 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.112278 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.215417 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.215469 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.215480 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.215502 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.215515 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.319060 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.319111 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.319128 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.319147 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.319160 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.355881 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:34 crc kubenswrapper[4906]: E1123 06:50:34.356132 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.422496 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.422577 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.422602 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.422637 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.422661 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.527257 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.527318 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.527337 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.527364 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.527383 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.630653 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.630755 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.630772 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.630801 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.630824 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.734984 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.735055 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.735075 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.735107 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.735129 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.760853 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:34 crc kubenswrapper[4906]: E1123 06:50:34.761104 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:34 crc kubenswrapper[4906]: E1123 06:50:34.761199 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:50:50.761170376 +0000 UTC m=+66.274561709 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.838845 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.838938 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.838965 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.839003 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.839024 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.942993 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.943041 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.943055 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.943080 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:34 crc kubenswrapper[4906]: I1123 06:50:34.943093 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:34Z","lastTransitionTime":"2025-11-23T06:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.047065 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.047143 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.047168 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.047200 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.047224 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.151983 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.152040 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.152056 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.152083 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.152101 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.255324 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.255380 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.255393 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.255415 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.255428 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.357941 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.358017 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.358118 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:35 crc kubenswrapper[4906]: E1123 06:50:35.358219 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:35 crc kubenswrapper[4906]: E1123 06:50:35.358403 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.358524 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: E1123 06:50:35.358553 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.358574 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.358597 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.358626 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.358647 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.381014 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.402015 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.424751 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.448985 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.462183 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.462435 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.462581 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.462754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.462943 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.471870 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.505211 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.543215 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.566762 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.567342 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.567427 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.567448 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.567481 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.567502 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.583933 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.609150 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.636547 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.654932 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.671785 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.671847 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.671869 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.671899 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.671920 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.673590 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.692787 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.705468 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.716989 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.728803 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:35Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.775575 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.775963 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.776237 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.776463 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.776629 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.881118 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.881208 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.881236 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.881277 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.881303 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.985612 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.985672 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.985707 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.985724 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:35 crc kubenswrapper[4906]: I1123 06:50:35.985737 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:35Z","lastTransitionTime":"2025-11-23T06:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.089273 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.089427 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.089447 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.089479 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.089499 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.192962 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.193012 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.193025 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.193049 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.193066 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.297141 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.297230 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.297250 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.297282 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.297302 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.355766 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:36 crc kubenswrapper[4906]: E1123 06:50:36.356908 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.402334 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.402469 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.402497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.402530 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.402556 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.507648 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.507744 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.507764 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.507796 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.507818 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.611645 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.611756 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.611779 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.611808 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.611829 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.715098 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.715175 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.715204 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.715243 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.715274 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.818623 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.818717 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.818744 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.818778 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.818804 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.922986 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.923060 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.923080 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.923113 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:36 crc kubenswrapper[4906]: I1123 06:50:36.923134 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:36Z","lastTransitionTime":"2025-11-23T06:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.026725 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.026791 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.026819 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.026856 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.026882 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.092592 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.092859 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.092956 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:51:09.092925822 +0000 UTC m=+84.606317165 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.093032 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.093073 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.093130 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093182 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093245 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093291 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093328 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093349 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093304 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:51:09.093274011 +0000 UTC m=+84.606665344 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093533 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093587 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093615 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093538 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:51:09.093470546 +0000 UTC m=+84.606861859 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093704 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:51:09.09366728 +0000 UTC m=+84.607058603 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.093745 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:51:09.093723872 +0000 UTC m=+84.607115225 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.130864 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.130954 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.131006 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.131032 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.131083 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.236571 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.236776 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.237313 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.237638 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.237989 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.341874 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.341934 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.341952 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.341974 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.341994 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.356131 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.356305 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.356588 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.356746 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.357156 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:37 crc kubenswrapper[4906]: E1123 06:50:37.357406 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.447165 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.447258 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.447290 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.447333 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.447356 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.550287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.550384 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.550410 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.550446 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.550476 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.654413 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.654485 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.654505 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.654531 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.654551 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.756927 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.756979 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.756999 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.757024 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.757045 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.859635 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.859730 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.859750 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.859769 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.859803 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.962351 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.962429 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.962448 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.962484 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:37 crc kubenswrapper[4906]: I1123 06:50:37.962507 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:37Z","lastTransitionTime":"2025-11-23T06:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.065864 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.065929 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.065940 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.065956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.065968 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.169144 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.169232 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.169252 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.169284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.169305 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.272974 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.273041 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.273054 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.273075 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.273089 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.355980 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:38 crc kubenswrapper[4906]: E1123 06:50:38.356215 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.377118 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.377176 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.377192 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.377217 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.377237 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.480911 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.480974 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.480991 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.481017 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.481035 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.584354 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.584403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.584421 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.584443 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.584460 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.688187 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.688260 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.688283 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.688314 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.688333 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.792409 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.792490 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.792514 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.792547 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.792571 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.896619 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.896730 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.896757 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.896790 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:38 crc kubenswrapper[4906]: I1123 06:50:38.896813 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:38Z","lastTransitionTime":"2025-11-23T06:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.000279 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.000363 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.000387 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.000424 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.000453 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.104557 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.104650 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.104712 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.104754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.104782 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.208614 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.208721 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.208756 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.208794 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.208820 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.312777 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.312876 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.312901 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.312938 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.312963 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.356070 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.356070 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.356081 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:39 crc kubenswrapper[4906]: E1123 06:50:39.356490 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:39 crc kubenswrapper[4906]: E1123 06:50:39.356627 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:39 crc kubenswrapper[4906]: E1123 06:50:39.356782 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.416785 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.416862 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.416886 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.416923 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.416949 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.519835 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.519901 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.519918 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.520375 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.520432 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.625084 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.625154 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.625181 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.625216 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.625242 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.729320 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.729446 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.729523 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.729564 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.729627 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.833254 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.833328 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.833353 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.833385 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.833409 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.936603 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.936665 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.936715 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.936746 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:39 crc kubenswrapper[4906]: I1123 06:50:39.936764 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:39Z","lastTransitionTime":"2025-11-23T06:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.040251 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.040346 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.040366 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.040402 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.040424 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.066717 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.083463 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.112314 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.144591 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.144647 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.144662 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.144707 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.144723 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.148191 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.178987 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.198182 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.223459 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.247936 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.248014 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.248035 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.248070 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.248089 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.254017 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.271084 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.289646 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.302313 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.317958 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.330180 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.346545 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.351702 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.351734 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.351747 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.351771 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.351787 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.356290 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:40 crc kubenswrapper[4906]: E1123 06:50:40.356454 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.361845 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.376149 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.402816 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.417512 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.433298 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:40Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.454888 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.454942 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.454955 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.454977 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.454993 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.557975 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.558046 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.558073 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.558102 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.558123 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.662244 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.662306 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.662321 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.662349 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.662364 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.766011 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.766061 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.766075 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.766094 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.766105 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.869504 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.869576 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.869604 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.869638 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.869655 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.973873 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.973960 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.974005 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.974048 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:40 crc kubenswrapper[4906]: I1123 06:50:40.974078 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:40Z","lastTransitionTime":"2025-11-23T06:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.078047 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.078118 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.078137 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.078169 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.078187 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.181456 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.181542 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.181570 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.181608 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.181637 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.285287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.285345 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.285360 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.285381 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.285395 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.355830 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.355909 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.355935 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.356013 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.356204 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.356248 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.388899 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.388963 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.388986 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.389020 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.389042 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.491978 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.492023 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.492036 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.492055 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.492064 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.595954 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.596047 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.596073 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.596109 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.596130 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.692185 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.692276 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.692298 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.692333 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.692357 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.716282 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:41Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.721658 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.721934 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.722106 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.722287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.722437 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.744289 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:41Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.750535 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.750787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.750951 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.751117 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.751256 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.772893 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:41Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.777835 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.777908 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.777927 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.777957 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.777976 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.798898 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:41Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.803945 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.803997 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.804016 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.804045 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.804064 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.824617 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:41Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:41 crc kubenswrapper[4906]: E1123 06:50:41.824823 4906 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.827051 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.827100 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.827109 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.827126 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.827141 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.930420 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.930496 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.930516 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.930549 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:41 crc kubenswrapper[4906]: I1123 06:50:41.930574 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:41Z","lastTransitionTime":"2025-11-23T06:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.034062 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.034123 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.034135 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.034156 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.034171 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.137000 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.137078 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.137099 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.137124 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.137142 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.240744 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.240815 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.240834 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.240866 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.240890 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.344945 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.345033 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.345058 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.345086 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.345108 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.355629 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:42 crc kubenswrapper[4906]: E1123 06:50:42.355923 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.448357 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.448425 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.448446 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.448476 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.448499 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.551918 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.551945 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.551955 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.551971 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.551981 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.655865 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.655942 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.655956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.655996 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.656013 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.759604 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.759704 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.759726 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.759757 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.759779 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.862266 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.862319 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.862337 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.862361 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.862379 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.965573 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.965645 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.965666 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.965723 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:42 crc kubenswrapper[4906]: I1123 06:50:42.965744 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:42Z","lastTransitionTime":"2025-11-23T06:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.069072 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.069142 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.069163 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.069190 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.069212 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.172124 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.172191 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.172211 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.172239 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.172259 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.276216 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.276289 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.276303 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.276322 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.276355 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.355813 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.355874 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.355823 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:43 crc kubenswrapper[4906]: E1123 06:50:43.356330 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:43 crc kubenswrapper[4906]: E1123 06:50:43.356448 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:43 crc kubenswrapper[4906]: E1123 06:50:43.356559 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.356755 4906 scope.go:117] "RemoveContainer" containerID="a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3" Nov 23 06:50:43 crc kubenswrapper[4906]: E1123 06:50:43.356918 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.379569 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.379651 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.379670 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.379738 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.379760 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.482491 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.482540 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.482554 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.482573 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.482586 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.585764 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.585805 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.585813 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.585826 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.585836 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.689048 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.689108 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.689127 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.689153 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.689172 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.792600 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.792671 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.792720 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.792746 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.792764 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.896174 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.896229 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.896247 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.896271 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.896286 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.999261 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.999313 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.999329 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.999356 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:43 crc kubenswrapper[4906]: I1123 06:50:43.999375 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:43Z","lastTransitionTime":"2025-11-23T06:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.102415 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.102495 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.102517 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.102550 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.102574 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.205107 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.205158 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.205171 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.205189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.205204 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.307465 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.307512 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.307523 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.307541 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.307553 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.356618 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:44 crc kubenswrapper[4906]: E1123 06:50:44.356916 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.410629 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.410773 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.410787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.410844 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.410861 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.513843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.513936 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.513950 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.513981 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.513995 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.617401 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.617497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.617520 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.617556 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.617579 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.719997 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.720053 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.720062 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.720079 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.720090 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.822510 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.822548 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.822560 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.822578 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.822590 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.925777 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.925843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.925861 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.925882 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:44 crc kubenswrapper[4906]: I1123 06:50:44.925897 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:44Z","lastTransitionTime":"2025-11-23T06:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.028724 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.028765 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.028773 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.028788 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.028800 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.131469 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.131516 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.131531 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.131564 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.131579 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.235139 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.235189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.235200 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.235215 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.235226 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.337983 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.338048 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.338063 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.338087 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.338101 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.355935 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.355975 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.355989 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:45 crc kubenswrapper[4906]: E1123 06:50:45.356075 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:45 crc kubenswrapper[4906]: E1123 06:50:45.356150 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:45 crc kubenswrapper[4906]: E1123 06:50:45.356228 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.391036 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.409567 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.426320 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.442207 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.442253 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.442267 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.442284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.442297 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.442470 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.459639 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.477500 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.515033 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.543146 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.545526 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.545595 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.545621 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.545656 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.545718 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.565517 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.588389 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.613944 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.638668 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.648984 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.649098 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.649122 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.649159 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.649180 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.660810 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.683419 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.698602 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.715449 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.739185 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.752863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.753255 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.753437 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.753618 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.753834 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.759420 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:45Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.856550 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.856608 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.856627 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.856657 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.856675 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.960628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.960760 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.960782 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.960816 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:45 crc kubenswrapper[4906]: I1123 06:50:45.960837 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:45Z","lastTransitionTime":"2025-11-23T06:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.064676 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.064746 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.064756 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.064776 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.064792 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.167844 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.167913 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.167930 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.167951 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.167966 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.272322 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.272395 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.272410 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.272457 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.272470 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.355901 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:46 crc kubenswrapper[4906]: E1123 06:50:46.356096 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.376196 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.376245 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.376258 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.376292 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.376307 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.479636 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.479703 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.479714 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.479733 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.479743 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.583787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.584342 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.584531 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.584716 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.584881 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.690054 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.690117 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.690133 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.690162 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.690185 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.800004 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.800080 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.800107 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.800150 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.800188 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.904342 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.904430 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.904456 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.904485 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:46 crc kubenswrapper[4906]: I1123 06:50:46.904506 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:46Z","lastTransitionTime":"2025-11-23T06:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.008797 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.009115 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.009128 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.009146 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.009160 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.111860 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.111935 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.111954 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.111984 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.112003 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.215320 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.215394 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.215412 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.215444 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.215467 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.318614 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.318715 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.318736 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.318765 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.318780 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.356092 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.356194 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:47 crc kubenswrapper[4906]: E1123 06:50:47.356337 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.356414 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:47 crc kubenswrapper[4906]: E1123 06:50:47.356606 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:47 crc kubenswrapper[4906]: E1123 06:50:47.356792 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.421358 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.421455 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.421476 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.421535 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.421556 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.525124 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.525202 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.525226 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.525259 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.525282 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.628822 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.628914 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.628935 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.628969 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.628988 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.733557 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.733718 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.733755 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.733792 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.733816 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.837429 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.837490 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.837505 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.837528 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.837543 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.940783 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.940826 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.940835 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.940850 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:47 crc kubenswrapper[4906]: I1123 06:50:47.940861 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:47Z","lastTransitionTime":"2025-11-23T06:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.044468 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.044551 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.044562 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.044606 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.044619 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.147053 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.147113 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.147125 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.147146 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.147159 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.250844 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.250924 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.250945 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.250977 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.250996 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.355371 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.355465 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.355483 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.355510 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.355524 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.355574 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:48 crc kubenswrapper[4906]: E1123 06:50:48.355907 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.458555 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.458639 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.458668 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.458746 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.458772 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.561654 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.561773 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.561791 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.561815 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.561833 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.664298 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.664385 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.664423 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.664457 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.664480 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.768009 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.768114 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.768145 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.768175 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.768201 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.871389 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.871435 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.871447 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.871463 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.871475 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.974862 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.974918 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.974929 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.974996 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:48 crc kubenswrapper[4906]: I1123 06:50:48.975021 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:48Z","lastTransitionTime":"2025-11-23T06:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.078444 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.078494 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.078521 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.078542 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.078556 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.182183 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.182231 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.182241 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.182263 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.182276 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.286338 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.286395 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.286436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.286455 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.286469 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.356056 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.356056 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:49 crc kubenswrapper[4906]: E1123 06:50:49.356240 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.356326 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:49 crc kubenswrapper[4906]: E1123 06:50:49.356445 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:49 crc kubenswrapper[4906]: E1123 06:50:49.356863 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.389831 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.389892 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.389909 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.389934 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.389952 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.493489 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.493814 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.493887 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.494035 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.494121 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.597171 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.597245 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.597267 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.597293 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.597313 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.700126 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.700171 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.700183 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.700199 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.700211 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.803351 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.803395 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.803404 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.803421 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.803429 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.905641 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.905960 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.906048 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.906150 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:49 crc kubenswrapper[4906]: I1123 06:50:49.906165 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:49Z","lastTransitionTime":"2025-11-23T06:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.008599 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.008672 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.008722 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.008752 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.008807 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.111672 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.111760 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.111778 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.111809 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.111828 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.214475 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.214524 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.214536 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.214554 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.214565 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.318008 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.318045 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.318067 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.318087 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.318102 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.356115 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:50 crc kubenswrapper[4906]: E1123 06:50:50.356256 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.420803 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.420845 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.420861 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.420882 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.420894 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.523815 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.523866 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.523877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.523893 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.523905 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.626553 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.626624 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.626649 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.626714 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.626742 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.730131 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.730184 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.730201 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.730229 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.730247 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.833016 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.833067 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.833078 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.833094 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.833112 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.852167 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:50 crc kubenswrapper[4906]: E1123 06:50:50.852286 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:50 crc kubenswrapper[4906]: E1123 06:50:50.852346 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:51:22.852329771 +0000 UTC m=+98.365721074 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.936718 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.936787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.936811 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.936841 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:50 crc kubenswrapper[4906]: I1123 06:50:50.936864 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:50Z","lastTransitionTime":"2025-11-23T06:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.040023 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.040077 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.040089 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.040108 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.040121 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.142984 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.143055 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.143077 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.143100 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.143118 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.246713 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.246747 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.246758 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.246773 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.246784 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.351404 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.351455 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.351466 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.351484 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.351499 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.355928 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:51 crc kubenswrapper[4906]: E1123 06:50:51.356093 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.356286 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:51 crc kubenswrapper[4906]: E1123 06:50:51.356348 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.356558 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:51 crc kubenswrapper[4906]: E1123 06:50:51.356632 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.454617 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.454709 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.454729 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.454752 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.454766 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.558829 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.558863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.558872 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.558888 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.558898 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.661782 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.661834 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.661845 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.661863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.661874 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.765089 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.765125 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.765135 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.765151 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.765160 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.867666 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.867843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.867882 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.867919 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.867940 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.971219 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.971267 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.971280 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.971295 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:51 crc kubenswrapper[4906]: I1123 06:50:51.971306 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:51Z","lastTransitionTime":"2025-11-23T06:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.073877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.073926 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.073942 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.073965 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.073981 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.143195 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.143227 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.143236 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.143957 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.143988 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: E1123 06:50:52.163080 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.167432 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.167485 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.167506 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.167529 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.167546 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: E1123 06:50:52.187537 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.193246 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.193285 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.193296 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.193315 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.193327 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: E1123 06:50:52.211101 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.215382 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.215415 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.215427 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.215440 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.215450 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: E1123 06:50:52.232061 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.235790 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.235837 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.235856 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.235878 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.235895 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: E1123 06:50:52.252080 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: E1123 06:50:52.252228 4906 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.254247 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.254271 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.254281 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.254295 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.254307 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.355592 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:52 crc kubenswrapper[4906]: E1123 06:50:52.355974 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.358372 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.358491 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.358609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.358735 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.358865 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.463184 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.463220 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.463230 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.463247 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.463260 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.565461 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.565498 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.565507 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.565520 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.565529 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.668043 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.668100 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.668111 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.668134 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.668146 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.777340 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.777414 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.777426 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.777445 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.777457 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.863980 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/0.log" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.864077 4906 generic.go:334] "Generic (PLEG): container finished" podID="4f75f165-ef56-40be-bfd4-1843de92b356" containerID="2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5" exitCode=1 Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.864135 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerDied","Data":"2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.864835 4906 scope.go:117] "RemoveContainer" containerID="2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.880208 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.880272 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.880292 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.880323 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.880342 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.883751 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.897757 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.917936 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.935331 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.955631 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.970440 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.984107 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.984644 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.984701 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.984712 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.984756 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.984765 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:52Z","lastTransitionTime":"2025-11-23T06:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:52 crc kubenswrapper[4906]: I1123 06:50:52.997060 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:52Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.015189 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.034957 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.051795 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.065099 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.086699 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.086957 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.087027 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.087118 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.086891 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.087183 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.104928 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.119375 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.136094 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.154403 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.169277 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.190202 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.190323 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.190386 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.190448 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.190507 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.293779 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.293848 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.293872 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.293906 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.293928 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.355756 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.355785 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:53 crc kubenswrapper[4906]: E1123 06:50:53.355902 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:53 crc kubenswrapper[4906]: E1123 06:50:53.356050 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.356087 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:53 crc kubenswrapper[4906]: E1123 06:50:53.356462 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.396984 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.397030 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.397040 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.397058 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.397071 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.499532 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.499710 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.499754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.499786 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.499811 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.603246 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.603310 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.603328 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.603353 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.603371 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.706711 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.707190 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.707342 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.707513 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.707663 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.811358 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.811413 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.811428 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.811453 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.811469 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.870013 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/0.log" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.870082 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerStarted","Data":"9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.886799 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.898332 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.913111 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.915052 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.915098 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.915111 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.915135 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.915148 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:53Z","lastTransitionTime":"2025-11-23T06:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.929740 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.941334 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.956617 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.967621 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.978068 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.987864 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:53 crc kubenswrapper[4906]: I1123 06:50:53.999846 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.010309 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.018337 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.018386 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.018401 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.018422 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.018437 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.022433 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.036948 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.055547 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.079880 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.098926 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.113891 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.122215 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.122255 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.122265 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.122283 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.122292 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.123750 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:54Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.224768 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.224822 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.224835 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.224853 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.224867 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.327459 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.327541 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.327559 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.327588 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.327609 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.356374 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:54 crc kubenswrapper[4906]: E1123 06:50:54.356581 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.430506 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.430576 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.430598 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.430627 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.430652 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.533019 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.533093 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.533111 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.533140 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.533159 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.636458 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.636500 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.636508 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.636527 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.636538 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.739452 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.739486 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.739494 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.739509 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.739518 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.843234 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.843316 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.843339 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.843376 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.843394 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.946117 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.946182 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.946206 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.946237 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:54 crc kubenswrapper[4906]: I1123 06:50:54.946258 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:54Z","lastTransitionTime":"2025-11-23T06:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.049191 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.049250 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.049277 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.049311 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.049333 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.151771 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.151861 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.151887 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.151920 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.151945 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.255032 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.255073 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.255082 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.255097 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.255107 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.355896 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:55 crc kubenswrapper[4906]: E1123 06:50:55.356033 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.356101 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.356482 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:55 crc kubenswrapper[4906]: E1123 06:50:55.356635 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:55 crc kubenswrapper[4906]: E1123 06:50:55.356913 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.360324 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.360357 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.360368 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.360384 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.360397 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.367588 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.371940 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.384368 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.397471 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.413189 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.428372 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.445387 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.459928 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.463250 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.463270 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.463279 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.463292 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.463302 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.471542 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.499977 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.525275 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.537512 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.548787 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.564959 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.568261 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.568306 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.568323 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.568346 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.568363 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.579811 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.592022 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.607440 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.622775 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.635132 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:55Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.671128 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.671191 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.671209 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.671233 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.671253 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.773881 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.773917 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.773928 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.773940 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.773950 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.876319 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.876761 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.876772 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.876786 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.876795 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.979839 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.979893 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.979909 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.979932 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:55 crc kubenswrapper[4906]: I1123 06:50:55.979950 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:55Z","lastTransitionTime":"2025-11-23T06:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.083155 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.083214 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.083233 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.083262 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.083280 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.185595 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.185652 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.185675 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.185739 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.185763 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.288515 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.288558 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.288570 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.288588 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.288603 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.356163 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:56 crc kubenswrapper[4906]: E1123 06:50:56.356843 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.357237 4906 scope.go:117] "RemoveContainer" containerID="a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.391523 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.391565 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.391579 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.391600 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.391609 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.494571 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.494627 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.494640 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.494658 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.494672 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.597270 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.597331 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.597350 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.597375 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.597394 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.700499 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.700596 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.700614 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.700641 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.700659 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.803865 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.803906 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.803921 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.803944 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.803961 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.883167 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/2.log" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.886155 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.886890 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.906752 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.906813 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.906829 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.906855 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.906871 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:56Z","lastTransitionTime":"2025-11-23T06:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.913457 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:56Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.930579 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:56Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.949475 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:56Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.968869 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:56Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:56 crc kubenswrapper[4906]: I1123 06:50:56.980292 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96a59f62-1019-4fff-bda1-643e107a433e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0128e3b7c58d98ebc49b75ecad88f2641115aff47aaef29a2a6bba6795c2b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:56Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.002951 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.009574 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.009598 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.009609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.009627 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.009638 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.014636 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.025944 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.037937 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.051007 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.063063 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.083514 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.104877 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.111831 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.111861 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.111870 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.111887 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.111898 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.118650 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.133178 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.158808 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.178243 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.195427 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.206913 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.214785 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.214840 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.214853 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.214873 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.214886 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.317455 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.317500 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.317513 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.317532 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.317544 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.355830 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.355830 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:57 crc kubenswrapper[4906]: E1123 06:50:57.355969 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.356060 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:57 crc kubenswrapper[4906]: E1123 06:50:57.356159 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:57 crc kubenswrapper[4906]: E1123 06:50:57.356322 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.439070 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.439121 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.439134 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.439151 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.439164 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.541516 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.541558 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.541567 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.541581 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.541590 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.645052 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.645106 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.645115 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.645137 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.645152 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.747666 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.747716 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.747725 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.747738 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.747747 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.861357 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.861410 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.861422 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.861439 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.861453 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.892445 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/3.log" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.893830 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/2.log" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.897840 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" exitCode=1 Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.897892 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.897940 4906 scope.go:117] "RemoveContainer" containerID="a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.899571 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 06:50:57 crc kubenswrapper[4906]: E1123 06:50:57.900005 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.946878 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a81f83ba755aeff9e51f0865d83a6ca4198d224d833a95199bf0b28e5c27f4a3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:30Z\\\",\\\"message\\\":\\\"andler 8\\\\nI1123 06:50:30.432195 6544 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:50:30.432245 6544 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:50:30.432301 6544 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:50:30.432352 6544 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:30.432368 6544 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:30.432390 6544 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:30.432626 6544 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432747 6544 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1123 06:50:30.432814 6544 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:30.432849 6544 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:30.432903 6544 factory.go:656] Stopping watch factory\\\\nI1123 06:50:30.432953 6544 ovnkube.go:599] Stopped ovnkube\\\\nI1123 06:50:30.432986 6544 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:30.432997 6544 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:30.433007 6544 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 06:50:30.433083 6544 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:57Z\\\",\\\"message\\\":\\\" 06:50:57.404809 6869 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:57.404827 6869 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:57.404834 6869 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:57.404859 6869 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:57.404893 6869 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:57.404900 6869 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:57.404942 6869 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:57.405528 6869 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:57.405560 6869 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:57.405564 6869 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:57.405576 6869 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:57.405601 6869 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:57.405639 6869 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.964044 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96a59f62-1019-4fff-bda1-643e107a433e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0128e3b7c58d98ebc49b75ecad88f2641115aff47aaef29a2a6bba6795c2b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.965224 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.965290 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.965302 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.965329 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.965345 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:57Z","lastTransitionTime":"2025-11-23T06:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:57 crc kubenswrapper[4906]: I1123 06:50:57.990363 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:57Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.007800 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.029032 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.047140 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.068834 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.068878 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.068890 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.068911 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.068926 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.068816 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.090496 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.110043 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.129188 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.147142 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.167273 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.171825 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.171853 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.171864 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.171880 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.171892 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.187372 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.208569 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.227268 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.246581 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.263838 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.274586 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.274628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.274637 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.274656 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.274665 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.277329 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.296535 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.356331 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:50:58 crc kubenswrapper[4906]: E1123 06:50:58.356589 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.377062 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.377132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.377158 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.377194 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.377220 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.480030 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.480387 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.480453 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.480535 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.480872 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.583511 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.583548 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.583557 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.583573 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.583583 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.687364 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.687420 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.687436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.687456 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.687473 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.790608 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.790674 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.790721 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.790749 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.790766 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.900624 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.900665 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.900697 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.900716 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.900731 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:58Z","lastTransitionTime":"2025-11-23T06:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.905652 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/3.log" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.909790 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 06:50:58 crc kubenswrapper[4906]: E1123 06:50:58.910021 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.929267 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.949509 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.964418 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:58 crc kubenswrapper[4906]: I1123 06:50:58.978277 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.004100 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.004155 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.004182 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.004325 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.004353 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.004362 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.021307 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.034928 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.051104 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.061798 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.074064 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.091769 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.104331 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.110219 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.110285 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.110307 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.110352 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.110374 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.129406 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:57Z\\\",\\\"message\\\":\\\" 06:50:57.404809 6869 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:57.404827 6869 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:57.404834 6869 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:57.404859 6869 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:57.404893 6869 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:57.404900 6869 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:57.404942 6869 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:57.405528 6869 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:57.405560 6869 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:57.405564 6869 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:57.405576 6869 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:57.405601 6869 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:57.405639 6869 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.146423 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96a59f62-1019-4fff-bda1-643e107a433e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0128e3b7c58d98ebc49b75ecad88f2641115aff47aaef29a2a6bba6795c2b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.177328 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.190867 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.207618 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.213255 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.213332 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.213347 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.213374 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.213391 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.224361 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.236001 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:50:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.316787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.316870 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.316895 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.316933 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.316957 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.355820 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.355895 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.355988 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:50:59 crc kubenswrapper[4906]: E1123 06:50:59.356058 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:50:59 crc kubenswrapper[4906]: E1123 06:50:59.356234 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:50:59 crc kubenswrapper[4906]: E1123 06:50:59.356331 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.420153 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.420222 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.420240 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.420267 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.420291 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.523567 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.523628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.523648 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.523710 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.523731 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.627634 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.627859 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.627914 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.627948 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.627969 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.732022 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.732132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.732157 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.732196 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.732216 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.835597 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.835654 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.835672 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.835730 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.835751 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.939028 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.939113 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.939141 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.939175 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:50:59 crc kubenswrapper[4906]: I1123 06:50:59.939198 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:50:59Z","lastTransitionTime":"2025-11-23T06:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.042575 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.042762 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.042813 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.042845 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.042867 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.147212 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.147269 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.147280 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.147300 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.147314 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.250232 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.250278 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.250287 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.250306 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.250317 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.353297 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.353457 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.353482 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.353511 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.353531 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.355871 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:00 crc kubenswrapper[4906]: E1123 06:51:00.356201 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.456636 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.456787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.456859 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.456892 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.456952 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.560540 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.560612 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.560638 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.560664 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.560722 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.664868 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.664950 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.664971 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.665022 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.665044 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.768922 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.769001 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.769027 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.769065 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.769096 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.873318 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.873386 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.873404 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.873436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.873456 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.977175 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.977243 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.977263 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.977295 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:00 crc kubenswrapper[4906]: I1123 06:51:00.977318 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:00Z","lastTransitionTime":"2025-11-23T06:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.080841 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.080897 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.080916 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.080938 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.080956 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.183967 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.184042 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.184063 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.184092 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.184113 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.287639 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.287750 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.287827 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.287861 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.287881 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.356545 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.356600 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.356734 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:01 crc kubenswrapper[4906]: E1123 06:51:01.356884 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:01 crc kubenswrapper[4906]: E1123 06:51:01.357033 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:01 crc kubenswrapper[4906]: E1123 06:51:01.357150 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.390466 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.390550 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.390568 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.390597 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.390615 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.494035 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.494092 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.494110 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.494139 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.494158 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.597947 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.598018 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.598035 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.598067 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.598089 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.701133 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.701216 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.701240 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.701266 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.701283 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.805289 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.805373 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.805398 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.805430 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.805452 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.909744 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.909812 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.909830 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.909857 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:01 crc kubenswrapper[4906]: I1123 06:51:01.909875 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:01Z","lastTransitionTime":"2025-11-23T06:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.014101 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.014189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.014210 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.014243 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.014265 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.117852 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.117913 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.117933 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.117965 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.117986 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.221589 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.221658 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.221711 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.221740 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.221762 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.325661 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.325761 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.325778 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.325803 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.325818 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.356524 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:02 crc kubenswrapper[4906]: E1123 06:51:02.356832 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.370564 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.370625 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.370646 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.370672 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.370746 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: E1123 06:51:02.387766 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:02Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.392480 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.392539 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.392557 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.392580 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.392596 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: E1123 06:51:02.416199 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:02Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.423158 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.423219 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.423265 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.423290 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.423308 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: E1123 06:51:02.437952 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:02Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.442754 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.442796 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.442811 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.442838 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.442852 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: E1123 06:51:02.458263 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:02Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.463328 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.463380 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.463391 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.463410 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.463423 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: E1123 06:51:02.477502 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:02Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:02 crc kubenswrapper[4906]: E1123 06:51:02.477620 4906 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.479978 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.480017 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.480030 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.480047 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.480059 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.583955 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.584048 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.584073 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.584106 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.584126 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.688172 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.688239 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.688256 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.688284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.688301 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.791544 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.791640 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.791663 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.791711 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.791728 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.895248 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.895311 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.895328 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.895353 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.895371 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.998529 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.998604 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.998628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.998658 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:02 crc kubenswrapper[4906]: I1123 06:51:02.998716 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:02Z","lastTransitionTime":"2025-11-23T06:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.101872 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.101950 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.101971 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.102003 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.102023 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.205502 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.205559 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.205578 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.205603 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.205620 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.317189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.317268 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.317285 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.317312 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.317324 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.356267 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.356267 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:03 crc kubenswrapper[4906]: E1123 06:51:03.356429 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.356517 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:03 crc kubenswrapper[4906]: E1123 06:51:03.356619 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:03 crc kubenswrapper[4906]: E1123 06:51:03.356707 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.419814 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.419856 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.419866 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.419882 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.419893 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.523110 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.523159 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.523196 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.523221 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.523233 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.625863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.625927 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.625945 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.625973 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.625993 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.729565 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.730172 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.730354 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.730587 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.730844 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.834666 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.834781 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.834806 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.834843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.834867 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.938434 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.938529 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.938560 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.938602 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:03 crc kubenswrapper[4906]: I1123 06:51:03.938630 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:03Z","lastTransitionTime":"2025-11-23T06:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.044066 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.044168 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.044189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.044221 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.044240 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.149216 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.149302 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.149321 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.149352 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.149369 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.253037 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.253107 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.253133 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.253161 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.253184 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.355714 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:04 crc kubenswrapper[4906]: E1123 06:51:04.355964 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.356293 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.356347 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.356369 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.356397 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.356417 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.459774 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.459837 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.459854 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.459878 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.459897 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.563261 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.563328 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.563351 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.563380 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.563399 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.667805 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.667871 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.667889 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.667915 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.667934 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.771282 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.771361 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.771379 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.771417 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.771439 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.874832 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.874914 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.874924 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.874967 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.874978 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.977822 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.977901 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.977920 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.977947 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:04 crc kubenswrapper[4906]: I1123 06:51:04.977969 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:04Z","lastTransitionTime":"2025-11-23T06:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.081624 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.081733 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.081762 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.081792 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.081809 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.184636 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.184763 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.184786 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.184812 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.184829 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.289134 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.289199 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.289217 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.289249 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.289268 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.356267 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.356418 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:05 crc kubenswrapper[4906]: E1123 06:51:05.356512 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.356537 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:05 crc kubenswrapper[4906]: E1123 06:51:05.356785 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:05 crc kubenswrapper[4906]: E1123 06:51:05.356910 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.392313 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.392410 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.392436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.392474 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.392498 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.396884 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:57Z\\\",\\\"message\\\":\\\" 06:50:57.404809 6869 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:57.404827 6869 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:57.404834 6869 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:57.404859 6869 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:57.404893 6869 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:57.404900 6869 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:57.404942 6869 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:57.405528 6869 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:57.405560 6869 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:57.405564 6869 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:57.405576 6869 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:57.405601 6869 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:57.405639 6869 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.416110 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96a59f62-1019-4fff-bda1-643e107a433e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0128e3b7c58d98ebc49b75ecad88f2641115aff47aaef29a2a6bba6795c2b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.452462 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.478003 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.495908 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.496096 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.496127 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.496166 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.496191 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.503188 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.521670 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.544361 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.564218 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.580811 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.598579 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.600078 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.600184 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.600210 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.600240 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.600299 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.615358 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.630860 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.654227 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.674406 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.692338 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.703991 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.704037 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.704049 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.704101 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.704119 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.708189 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.726563 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.743234 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.762489 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:05Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.807400 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.807486 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.807505 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.807542 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.807564 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.910327 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.910855 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.911025 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.911178 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:05 crc kubenswrapper[4906]: I1123 06:51:05.911329 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:05Z","lastTransitionTime":"2025-11-23T06:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.015447 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.015511 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.015524 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.015548 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.015562 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.118748 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.119127 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.119288 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.119456 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.119604 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.223850 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.224161 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.224718 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.224886 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.225042 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.329417 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.329507 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.329526 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.329560 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.329578 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.356223 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:06 crc kubenswrapper[4906]: E1123 06:51:06.356650 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.433803 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.434321 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.434342 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.434372 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.434391 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.537519 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.537576 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.537586 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.537608 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.537620 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.643551 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.643632 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.643656 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.643728 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.643753 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.746996 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.747091 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.747117 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.747155 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.747183 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.851985 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.852084 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.852109 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.852159 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.852180 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.955577 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.955642 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.955664 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.955749 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:06 crc kubenswrapper[4906]: I1123 06:51:06.955775 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:06Z","lastTransitionTime":"2025-11-23T06:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.059050 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.059112 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.059124 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.059148 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.059164 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.161321 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.161374 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.161386 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.161406 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.161421 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.264861 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.264931 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.264953 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.264986 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.265004 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.356113 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.356176 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.356150 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:07 crc kubenswrapper[4906]: E1123 06:51:07.356397 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:07 crc kubenswrapper[4906]: E1123 06:51:07.356631 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:07 crc kubenswrapper[4906]: E1123 06:51:07.356801 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.367463 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.367509 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.367527 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.367555 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.367573 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.472097 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.472167 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.472180 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.472204 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.472220 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.576032 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.576079 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.576088 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.576124 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.576134 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.680566 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.680667 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.680739 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.680792 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.680820 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.784422 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.784761 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.784922 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.785199 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.785337 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.888607 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.889064 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.889220 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.889375 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.889517 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.992359 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.992406 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.992416 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.992435 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:07 crc kubenswrapper[4906]: I1123 06:51:07.992447 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:07Z","lastTransitionTime":"2025-11-23T06:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.095671 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.095792 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.095817 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.095852 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.095882 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.199570 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.199795 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.199824 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.199854 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.199873 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.303549 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.303600 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.303611 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.303628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.303639 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.355866 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:08 crc kubenswrapper[4906]: E1123 06:51:08.356240 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.408513 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.408573 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.408592 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.408622 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.408644 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.512618 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.512760 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.512783 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.512814 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.512832 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.616402 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.616471 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.616488 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.616513 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.616536 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.720989 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.721079 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.721103 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.721146 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.721173 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.825549 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.825621 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.825640 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.825729 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.825763 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.929352 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.929438 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.929455 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.929486 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:08 crc kubenswrapper[4906]: I1123 06:51:08.929506 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:08Z","lastTransitionTime":"2025-11-23T06:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.032554 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.032625 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.032648 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.032676 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.032730 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.137181 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.137261 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.137295 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.137341 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.137368 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.182865 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.183555 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.183729 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.183813 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.184153 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.184108687 +0000 UTC m=+148.697500020 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.184281 4906 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.184366 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.184348013 +0000 UTC m=+148.697739346 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.184820 4906 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.184898 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.184881527 +0000 UTC m=+148.698272870 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.185204 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.185240 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.185261 4906 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.185318 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.185302987 +0000 UTC m=+148.698694330 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.183981 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.191124 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.191306 4906 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.191335 4906 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.191460 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.191420479 +0000 UTC m=+148.704811822 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.241805 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.241900 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.241923 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.241956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.241978 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.346664 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.346780 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.346800 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.346830 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.346849 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.356088 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.356195 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.356290 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.356455 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.356455 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:09 crc kubenswrapper[4906]: E1123 06:51:09.357067 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.451060 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.451115 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.451132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.451160 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.451180 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.554939 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.555009 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.555028 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.555054 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.555075 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.659078 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.659170 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.659190 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.659224 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.659248 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.763381 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.763478 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.763498 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.763536 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.763557 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.867951 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.868055 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.868076 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.868108 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.868129 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.971875 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.971956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.971980 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.972012 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:09 crc kubenswrapper[4906]: I1123 06:51:09.972032 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:09Z","lastTransitionTime":"2025-11-23T06:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.075543 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.075609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.075627 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.075652 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.075671 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.180406 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.180479 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.180496 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.180525 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.180546 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.284091 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.284158 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.284177 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.284206 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.284227 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.355857 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:10 crc kubenswrapper[4906]: E1123 06:51:10.356331 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.388913 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.388979 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.388997 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.389023 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.389041 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.493144 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.493222 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.493240 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.493301 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.493321 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.597352 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.597443 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.597463 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.597497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.597522 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.701547 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.702307 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.702335 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.702368 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.702388 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.806028 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.806125 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.806151 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.806187 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.806216 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.909327 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.909421 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.909440 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.909471 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:10 crc kubenswrapper[4906]: I1123 06:51:10.909491 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:10Z","lastTransitionTime":"2025-11-23T06:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.012560 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.012615 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.012627 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.012648 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.012661 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.116660 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.116776 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.116796 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.116826 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.116845 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.220365 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.220445 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.220465 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.220501 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.220529 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.324851 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.324932 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.324952 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.324982 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.325003 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.355525 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.355617 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.355625 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:11 crc kubenswrapper[4906]: E1123 06:51:11.355805 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:11 crc kubenswrapper[4906]: E1123 06:51:11.355958 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:11 crc kubenswrapper[4906]: E1123 06:51:11.356154 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.428858 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.428944 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.428966 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.428995 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.429052 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.531996 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.532086 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.532112 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.532149 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.532174 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.635884 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.635973 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.635992 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.636023 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.636044 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.739315 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.739379 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.739402 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.739430 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.739450 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.843442 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.843499 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.843520 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.843551 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.843573 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.947411 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.947499 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.947531 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.947569 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:11 crc kubenswrapper[4906]: I1123 06:51:11.947597 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:11Z","lastTransitionTime":"2025-11-23T06:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.052156 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.052227 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.052250 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.052285 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.052313 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.156093 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.156183 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.156205 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.156246 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.156270 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.260531 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.260610 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.260629 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.260657 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.260713 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.356452 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.356673 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.358046 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.358398 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.364505 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.364550 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.364570 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.364599 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.364618 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.467950 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.468021 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.468040 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.468071 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.468095 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.552292 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.552360 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.552380 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.552409 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.552443 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.574615 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.580118 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.580173 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.580188 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.580214 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.580232 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.600331 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.605799 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.605870 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.605885 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.605930 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.605947 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.622912 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.628776 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.628836 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.628849 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.628872 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.628892 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.651141 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.656763 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.656833 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.656851 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.656888 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.656911 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.675821 4906 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:51:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9217c3d2-062e-44c4-9092-623fa669de19\\\",\\\"systemUUID\\\":\\\"9f63f9ce-8df2-4d0e-becf-5452e97f2c1a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:12 crc kubenswrapper[4906]: E1123 06:51:12.676094 4906 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.678731 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.678803 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.678821 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.678841 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.678856 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.781528 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.781595 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.781609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.781628 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.781642 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.885475 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.885563 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.885584 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.885617 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.885636 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.988305 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.988383 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.988403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.988440 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:12 crc kubenswrapper[4906]: I1123 06:51:12.988463 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:12Z","lastTransitionTime":"2025-11-23T06:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.092314 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.092385 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.092403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.092431 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.092456 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.205820 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.205889 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.205908 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.205934 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.205954 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.309997 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.310066 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.310087 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.310120 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.310143 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.356113 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.356150 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.356262 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:13 crc kubenswrapper[4906]: E1123 06:51:13.356370 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:13 crc kubenswrapper[4906]: E1123 06:51:13.356567 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:13 crc kubenswrapper[4906]: E1123 06:51:13.356658 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.413382 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.413448 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.413468 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.413497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.413517 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.517451 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.517609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.517640 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.517671 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.517797 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.621765 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.621839 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.621866 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.621899 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.621922 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.725217 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.725296 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.725318 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.725348 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.725368 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.830081 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.830156 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.830176 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.830207 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.830226 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.933563 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.933635 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.933659 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.933715 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:13 crc kubenswrapper[4906]: I1123 06:51:13.933738 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:13Z","lastTransitionTime":"2025-11-23T06:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.037787 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.037863 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.037882 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.037909 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.037935 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.142493 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.142751 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.142778 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.142854 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.142879 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.246619 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.246730 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.246751 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.246783 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.246803 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.350133 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.350206 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.350232 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.350266 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.350289 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.355664 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:14 crc kubenswrapper[4906]: E1123 06:51:14.355908 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.454023 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.454170 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.454189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.454225 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.454248 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.558872 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.558976 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.558997 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.559030 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.559050 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.662832 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.662890 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.662917 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.662940 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.662956 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.765998 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.766111 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.766138 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.766180 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.766212 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.869067 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.869139 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.869164 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.869196 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.869215 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.973866 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.973943 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.973962 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.973989 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:14 crc kubenswrapper[4906]: I1123 06:51:14.974012 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:14Z","lastTransitionTime":"2025-11-23T06:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.078109 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.078189 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.078211 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.078243 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.078263 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.182634 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.182743 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.182766 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.182802 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.182825 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.287305 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.287388 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.287417 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.287449 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.287473 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.356365 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.356582 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.356954 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:15 crc kubenswrapper[4906]: E1123 06:51:15.356951 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:15 crc kubenswrapper[4906]: E1123 06:51:15.357613 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:15 crc kubenswrapper[4906]: E1123 06:51:15.357731 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.377284 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec5c6569-64d4-4591-bbac-b70ebcca836f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d96410695caf0070f5a8e0c18c4fadf2bf610f63c62f8322ef1872dfc14e1766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d4j66\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2xgvr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.390646 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.390722 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.390742 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.390767 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.390785 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.399905 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.421726 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.442939 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lm9xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f75f165-ef56-40be-bfd4-1843de92b356\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:52Z\\\",\\\"message\\\":\\\"2025-11-23T06:50:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d\\\\n2025-11-23T06:50:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c74e0132-4338-4539-9914-92a874cca87d to /host/opt/cni/bin/\\\\n2025-11-23T06:50:07Z [verbose] multus-daemon started\\\\n2025-11-23T06:50:07Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:50:52Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmffv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lm9xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.472585 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3267a7a66e89ba701214490a78d419452a3036705e2d64c0314917712bb71756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.493441 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.493533 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.493552 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.493582 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.493602 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.508383 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d6f3b5-c353-4412-8f38-989b53ddef2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:50:57Z\\\",\\\"message\\\":\\\" 06:50:57.404809 6869 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:50:57.404827 6869 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:50:57.404834 6869 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:50:57.404859 6869 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:50:57.404893 6869 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1123 06:50:57.404900 6869 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1123 06:50:57.404942 6869 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1123 06:50:57.405528 6869 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:50:57.405560 6869 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:50:57.405564 6869 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:50:57.405576 6869 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1123 06:50:57.405582 6869 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:50:57.405601 6869 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:50:57.405639 6869 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hvsb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tkkm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.527101 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96a59f62-1019-4fff-bda1-643e107a433e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0128e3b7c58d98ebc49b75ecad88f2641115aff47aaef29a2a6bba6795c2b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59de12a2c3011b9ffb126f049b0fcb694b80f9acf24afa8a71cde39243839d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.556285 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6fd8e94-d5c4-4eaf-975a-e74fe976a104\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03d8b80fc63f6e013e64c57aaf1e35ebeeeab1772b5d8098aa3b2f99a65ea345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268a414844f1e25fd9420beee6fef59ff3421fdf562c59d52882c26d48df2a23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7165d8b3f507341bddab2b89d1bec1a972eeb90e90e9f5b06dbf6a0ceac4cfb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://167c2602ad65f58908ab0aad719db45dbe7ea34fd32babcce15984d72c3e23a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d53bdadbd42794654ac81d8ad2542b4da480bd6d7633429cc9aca282d237ab94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82ccfc29e9b3a416d89b0ba27b51c3b9ee33b868b3432b14ad57c4e0b208cc93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d3bb48b8b086dfe894d84a7e6bff86421f9ad40e641ccd9e905912243c3f6e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b537eb136b2f868d6485511334c6f66d5da524d24e3e7177d0763a17ada88e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.578104 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7c6fbaf-fa9c-492a-b472-c40d18ad5931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d91cb6eacb0e77316bb30847e5bcea1df59de2bd67d652403441e45b3e61e6a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3195d9fee563c3d5877ad11681e4759eea52c6871034e405c1c4edb21e4bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://179b66141cd3aa4a3bbdd52cbafcc6cdcf89da0491cef41cddc03d1b7bd6a853\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77785e50b8c1425d5f35868cff97bea1495c5a783d52eaf2ad09da67e2a239fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.597254 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8drn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn5t8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8drn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.597403 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.597436 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.597455 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.597483 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.597502 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.621784 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6mz65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e886ca-acf1-4bf0-94fd-69dd208d8da9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b43e108b9e95356dee0b25c7ed9450a43e3f0640365a74c68d5f38609756fada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901079812b30891863cfef21375bf2acc4184b45044c93be6ef2ff186d8c0a06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://468f7b2f8b3818a8e9303f6e047d2bbebd5e2d6a591ed239286b7e6446d1eb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30c439a733e32a221732a4dd825f599e4733352fa89f148450594e0d6a44e365\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c3c7b4727f63a7682388aeae5ceb06a86aca61354d28f6592921f3e7f6b2af1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8af1588f962fd4417fad79e3ff9cf2064e16527a2656a0b70e6b0fd4059c4b60\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://815b36888dc986aec851b310131c4b1671d5bb9686cd387cf89984ae0bd2de71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:50:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:50:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w7hrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6mz65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.646320 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://099a58ce157c3251279b32001a35c3cde7bf53c6e50c7089866ad4f33c9b1df5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9337027be88d28e9a670c2fa462bc771c03947ba535dce9e821c2d821f6b56d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.668853 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8b0a056-21ed-4d43-bde9-5ecff043aaa4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606692abb33c0f792e9d8273304de544da6d53e4469b87ea5b3076c9bf869060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa16b1dab92f4e3bc9519b21e8922bf1ef32e832e923f96f1d7692e00185f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjkzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6tlzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.694168 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c7a040b-fb78-4913-b589-5c080ee5cfc3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23e3086f8b42260bf616f2d720d3c1416168d738dbb0eb53e16e23ff12850c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4609908baa0e4016d236bfa79a7ff6f91f6b8e5647fd096e02593eb03c8e24df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1839a48aa568645b3f55230c26fa85ac7eeb5e06d61c1c98357daca50c2fd1e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba18b88d9ea449002cbfbac53d983c2a502bd33fb071e153f8f200093b7cff8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://323961e6dda20bfe86ea524ade9b479b4fb812648a232ba86cf22f87ca2658f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:50:04Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1123 06:49:48.915778 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:49:48.918729 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1326598051/tls.crt::/tmp/serving-cert-1326598051/tls.key\\\\\\\"\\\\nI1123 06:50:04.365695 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:50:04.368401 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:50:04.368423 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:50:04.368449 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:50:04.368455 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:50:04.375886 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:50:04.375910 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:50:04.375910 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:50:04.375919 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:50:04.375930 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:50:04.375936 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:50:04.375940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:50:04.375943 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:50:04.379981 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58b92842b1711f48b4aea0c697cf37e1447d8c5755e3b3ccfee6a8a18e59d094\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7c2d36d19f15e896d5c5846502a52f99020a978bba35e12fafc63956e70359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.704497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.704580 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.704600 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.704635 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.704654 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.717348 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d7fff5494552b5845a6e4d2c559e4f4d8a244ef6ce004a022ef753cb3848034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.737851 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.760917 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85067bcd-aa25-458c-83ce-32dcea5b182e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de28db8b64c311dce393e206cc77537838683069b63183c59c029217d4535dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8076ee1caf337cd6df0e7832229d44bf748c8aeafa848851855058c4343991ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a74c6cc8aa48c757964c74fe10401f90b564d1b533e97626213d03d5c75b8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:49:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f8e7a4a5ad213067d8587e6f5134b9d09ed330d8ce617c2a291468033062a8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:49:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:49:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:49:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.804290 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jv7sg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c881afa1-c693-450f-912c-cbb4c33bb04f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af210d765c625621b57e2ff4bed17725f4ac86d9889e4761c1f2ae93d78f6635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqvhh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jv7sg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.808177 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.808229 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.808247 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.808277 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.808299 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.824287 4906 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pd47s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"407b04ac-8bbb-4112-88bc-bbfc89e37f6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:50:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26d53d375fdf3241904bf3dcbe4245b40cb8a09b1c7878c91bc1fd9607d5164d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:50:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:50:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pd47s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:51:15Z is after 2025-08-24T17:21:41Z" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.912497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.912586 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.912612 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.912654 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:15 crc kubenswrapper[4906]: I1123 06:51:15.912715 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:15Z","lastTransitionTime":"2025-11-23T06:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.017260 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.017331 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.017355 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.017386 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.017408 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.120854 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.120915 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.120931 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.120956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.120973 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.224758 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.224836 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.224855 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.224885 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.224906 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.328796 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.328856 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.328874 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.328900 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.328920 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.356466 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:16 crc kubenswrapper[4906]: E1123 06:51:16.356741 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.433119 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.433179 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.433198 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.433261 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.433279 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.536254 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.536319 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.536339 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.536374 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.536398 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.639877 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.639956 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.639975 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.640007 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.640029 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.750284 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.750370 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.750392 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.750426 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.750447 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.853521 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.853594 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.853616 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.853647 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.853665 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.956749 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.956824 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.956843 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.956869 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:16 crc kubenswrapper[4906]: I1123 06:51:16.956892 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:16Z","lastTransitionTime":"2025-11-23T06:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.060906 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.060968 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.060983 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.061006 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.061024 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.164140 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.164194 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.164207 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.164227 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.164243 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.268309 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.268401 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.268425 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.268461 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.268486 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.356088 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.356317 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:17 crc kubenswrapper[4906]: E1123 06:51:17.356416 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.356539 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:17 crc kubenswrapper[4906]: E1123 06:51:17.356549 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:17 crc kubenswrapper[4906]: E1123 06:51:17.356636 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.370785 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.370848 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.370867 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.370894 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.370915 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.473825 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.473897 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.473919 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.473950 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.473970 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.577261 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.577619 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.577767 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.577873 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.577966 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.682375 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.682697 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.682798 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.682886 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.682963 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.786010 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.786643 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.786832 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.786967 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.787095 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.889960 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.890407 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.890568 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.890778 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.890930 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.994795 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.994890 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.994907 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.994934 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:17 crc kubenswrapper[4906]: I1123 06:51:17.994956 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:17Z","lastTransitionTime":"2025-11-23T06:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.099223 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.099330 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.099352 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.099380 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.099399 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.202428 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.202513 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.202539 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.202571 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.202595 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.305915 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.305975 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.305990 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.306008 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.306022 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.356161 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:18 crc kubenswrapper[4906]: E1123 06:51:18.356440 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.409535 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.409596 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.409609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.409630 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.409646 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.513781 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.513862 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.513879 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.513941 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.513959 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.617991 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.618061 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.618081 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.618115 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.618135 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.723160 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.723231 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.723264 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.723295 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.723319 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.827627 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.827715 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.827736 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.827797 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.827817 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.934163 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.934309 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.934330 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.934394 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:18 crc kubenswrapper[4906]: I1123 06:51:18.934414 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:18Z","lastTransitionTime":"2025-11-23T06:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.037939 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.038004 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.038021 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.038046 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.038063 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.141454 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.141549 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.141576 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.141609 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.141628 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.246015 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.246051 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.246060 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.246077 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.246086 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.354081 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.354122 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.354132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.354147 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.354157 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.355476 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.355548 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:19 crc kubenswrapper[4906]: E1123 06:51:19.355580 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:19 crc kubenswrapper[4906]: E1123 06:51:19.355704 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.355787 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:19 crc kubenswrapper[4906]: E1123 06:51:19.355864 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.457484 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.457549 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.457567 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.457599 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.457619 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.560614 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.560974 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.561102 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.561206 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.561304 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.664121 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.664183 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.664203 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.664230 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.664248 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.767697 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.767755 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.767774 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.767798 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.767814 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.870935 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.871025 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.871073 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.871108 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.871129 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.974129 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.974193 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.974212 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.974239 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:19 crc kubenswrapper[4906]: I1123 06:51:19.974257 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:19Z","lastTransitionTime":"2025-11-23T06:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.077389 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.077447 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.077468 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.077492 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.077509 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.181074 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.181117 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.181129 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.181144 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.181156 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.283733 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.283827 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.283855 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.283888 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.283914 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.355381 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:20 crc kubenswrapper[4906]: E1123 06:51:20.355508 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.387147 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.387183 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.387196 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.387214 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.387228 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.490530 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.490578 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.490593 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.490615 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.490638 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.594497 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.594538 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.594548 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.594563 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.594575 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.697132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.697192 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.697211 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.697233 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.697250 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.800231 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.800301 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.800319 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.800356 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.800381 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.903391 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.903437 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.903448 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.903464 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:20 crc kubenswrapper[4906]: I1123 06:51:20.903473 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:20Z","lastTransitionTime":"2025-11-23T06:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.007024 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.007080 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.007092 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.007113 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.007126 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.110339 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.110398 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.110408 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.110427 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.110439 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.213890 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.213969 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.213987 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.214014 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.214031 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.317981 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.318059 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.318086 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.318118 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.318140 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.356250 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.356260 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.356284 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:21 crc kubenswrapper[4906]: E1123 06:51:21.356521 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:21 crc kubenswrapper[4906]: E1123 06:51:21.356621 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:21 crc kubenswrapper[4906]: E1123 06:51:21.357157 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.422911 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.422994 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.423017 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.423048 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.423073 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.526992 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.527090 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.527116 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.527152 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.527176 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.631411 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.631484 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.631503 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.631531 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.631549 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.735199 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.735273 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.735291 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.735319 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.735339 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.838985 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.839050 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.839068 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.839095 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.839113 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.942513 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.942581 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.942600 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.942630 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:21 crc kubenswrapper[4906]: I1123 06:51:21.942649 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:21Z","lastTransitionTime":"2025-11-23T06:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.046600 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.046730 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.046758 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.046794 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.046819 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.150365 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.150431 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.150453 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.150479 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.150498 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.254448 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.254528 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.254546 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.254573 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.254592 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.355990 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:22 crc kubenswrapper[4906]: E1123 06:51:22.356206 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.358553 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.358648 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.358667 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.358720 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.358738 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.462445 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.462519 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.462537 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.462568 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.462592 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.566865 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.566923 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.566938 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.566960 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.566974 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.670500 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.670580 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.670605 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.670641 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.670666 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.773805 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.773882 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.773902 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.773930 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.773949 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.904939 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.905010 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.905028 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.905054 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.905073 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.907024 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.907132 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.907183 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.907210 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.907230 4906 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:51:22Z","lastTransitionTime":"2025-11-23T06:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.952967 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:22 crc kubenswrapper[4906]: E1123 06:51:22.953282 4906 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:51:22 crc kubenswrapper[4906]: E1123 06:51:22.953444 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs podName:98bb1da0-e8ce-4b13-8a4d-c30423c5baa8 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:26.953406915 +0000 UTC m=+162.466798258 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs") pod "network-metrics-daemon-8drn8" (UID: "98bb1da0-e8ce-4b13-8a4d-c30423c5baa8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.989934 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc"] Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.990665 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.995770 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.996173 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.996282 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 23 06:51:22 crc kubenswrapper[4906]: I1123 06:51:22.998889 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.014858 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podStartSLOduration=79.014772556 podStartE2EDuration="1m19.014772556s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.014336734 +0000 UTC m=+98.527728057" watchObservedRunningTime="2025-11-23 06:51:23.014772556 +0000 UTC m=+98.528163899" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.054525 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b338ad46-b672-4399-9957-da15a42f7384-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.054645 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b338ad46-b672-4399-9957-da15a42f7384-service-ca\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.054803 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b338ad46-b672-4399-9957-da15a42f7384-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.054852 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b338ad46-b672-4399-9957-da15a42f7384-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.054951 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b338ad46-b672-4399-9957-da15a42f7384-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.097332 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-lm9xt" podStartSLOduration=79.097296703 podStartE2EDuration="1m19.097296703s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.071373302 +0000 UTC m=+98.584764645" watchObservedRunningTime="2025-11-23 06:51:23.097296703 +0000 UTC m=+98.610688036" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.155724 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b338ad46-b672-4399-9957-da15a42f7384-service-ca\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.155835 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b338ad46-b672-4399-9957-da15a42f7384-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.155887 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b338ad46-b672-4399-9957-da15a42f7384-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.156016 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b338ad46-b672-4399-9957-da15a42f7384-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.156106 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b338ad46-b672-4399-9957-da15a42f7384-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.156213 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b338ad46-b672-4399-9957-da15a42f7384-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.156237 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b338ad46-b672-4399-9957-da15a42f7384-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.157511 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b338ad46-b672-4399-9957-da15a42f7384-service-ca\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.174346 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.174312465 podStartE2EDuration="28.174312465s" podCreationTimestamp="2025-11-23 06:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.17334433 +0000 UTC m=+98.686735643" watchObservedRunningTime="2025-11-23 06:51:23.174312465 +0000 UTC m=+98.687703808" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.174769 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b338ad46-b672-4399-9957-da15a42f7384-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.180877 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b338ad46-b672-4399-9957-da15a42f7384-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-rg8gc\" (UID: \"b338ad46-b672-4399-9957-da15a42f7384\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.248288 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.248242806 podStartE2EDuration="1m16.248242806s" podCreationTimestamp="2025-11-23 06:50:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.224852972 +0000 UTC m=+98.738244325" watchObservedRunningTime="2025-11-23 06:51:23.248242806 +0000 UTC m=+98.761634119" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.249026 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.249016937 podStartE2EDuration="1m19.249016937s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.247825835 +0000 UTC m=+98.761217168" watchObservedRunningTime="2025-11-23 06:51:23.249016937 +0000 UTC m=+98.762408260" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.314094 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.319276 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6tlzk" podStartSLOduration=78.319249341 podStartE2EDuration="1m18.319249341s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.31882251 +0000 UTC m=+98.832213853" watchObservedRunningTime="2025-11-23 06:51:23.319249341 +0000 UTC m=+98.832640654" Nov 23 06:51:23 crc kubenswrapper[4906]: W1123 06:51:23.335065 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb338ad46_b672_4399_9957_da15a42f7384.slice/crio-c00fed4827d71d64f0932948ff4e5dc1785ce31de6e3cecbe6dec95f1cf7417a WatchSource:0}: Error finding container c00fed4827d71d64f0932948ff4e5dc1785ce31de6e3cecbe6dec95f1cf7417a: Status 404 returned error can't find the container with id c00fed4827d71d64f0932948ff4e5dc1785ce31de6e3cecbe6dec95f1cf7417a Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.356306 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.356390 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:23 crc kubenswrapper[4906]: E1123 06:51:23.356473 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.356671 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:23 crc kubenswrapper[4906]: E1123 06:51:23.356676 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:23 crc kubenswrapper[4906]: E1123 06:51:23.357116 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.369633 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.369606873 podStartE2EDuration="1m19.369606873s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.342534192 +0000 UTC m=+98.855925515" watchObservedRunningTime="2025-11-23 06:51:23.369606873 +0000 UTC m=+98.882998186" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.431306 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-6mz65" podStartSLOduration=79.431280402 podStartE2EDuration="1m19.431280402s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.41288886 +0000 UTC m=+98.926280203" watchObservedRunningTime="2025-11-23 06:51:23.431280402 +0000 UTC m=+98.944671705" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.432264 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.432259048 podStartE2EDuration="43.432259048s" podCreationTimestamp="2025-11-23 06:50:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.430380578 +0000 UTC m=+98.943771891" watchObservedRunningTime="2025-11-23 06:51:23.432259048 +0000 UTC m=+98.945650351" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.447742 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-jv7sg" podStartSLOduration=79.447721794 podStartE2EDuration="1m19.447721794s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.447594881 +0000 UTC m=+98.960986174" watchObservedRunningTime="2025-11-23 06:51:23.447721794 +0000 UTC m=+98.961113097" Nov 23 06:51:23 crc kubenswrapper[4906]: I1123 06:51:23.467276 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-pd47s" podStartSLOduration=79.467254207 podStartE2EDuration="1m19.467254207s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:23.467077253 +0000 UTC m=+98.980468556" watchObservedRunningTime="2025-11-23 06:51:23.467254207 +0000 UTC m=+98.980645520" Nov 23 06:51:24 crc kubenswrapper[4906]: I1123 06:51:24.032222 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" event={"ID":"b338ad46-b672-4399-9957-da15a42f7384","Type":"ContainerStarted","Data":"950c1f0d742afce1a5746cdb35604d0a96463dd058e6671fb509eaa5c6c3ab5e"} Nov 23 06:51:24 crc kubenswrapper[4906]: I1123 06:51:24.032326 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" event={"ID":"b338ad46-b672-4399-9957-da15a42f7384","Type":"ContainerStarted","Data":"c00fed4827d71d64f0932948ff4e5dc1785ce31de6e3cecbe6dec95f1cf7417a"} Nov 23 06:51:24 crc kubenswrapper[4906]: I1123 06:51:24.060339 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rg8gc" podStartSLOduration=80.06031676 podStartE2EDuration="1m20.06031676s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:24.059433027 +0000 UTC m=+99.572824370" watchObservedRunningTime="2025-11-23 06:51:24.06031676 +0000 UTC m=+99.573708073" Nov 23 06:51:24 crc kubenswrapper[4906]: I1123 06:51:24.356283 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:24 crc kubenswrapper[4906]: E1123 06:51:24.356505 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:24 crc kubenswrapper[4906]: I1123 06:51:24.357538 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 06:51:24 crc kubenswrapper[4906]: E1123 06:51:24.357876 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tkkm2_openshift-ovn-kubernetes(92d6f3b5-c353-4412-8f38-989b53ddef2a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" Nov 23 06:51:25 crc kubenswrapper[4906]: I1123 06:51:25.357844 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:25 crc kubenswrapper[4906]: E1123 06:51:25.358165 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:25 crc kubenswrapper[4906]: I1123 06:51:25.358827 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:25 crc kubenswrapper[4906]: E1123 06:51:25.358981 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:25 crc kubenswrapper[4906]: I1123 06:51:25.359038 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:25 crc kubenswrapper[4906]: E1123 06:51:25.359292 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:26 crc kubenswrapper[4906]: I1123 06:51:26.355856 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:26 crc kubenswrapper[4906]: E1123 06:51:26.356039 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:27 crc kubenswrapper[4906]: I1123 06:51:27.356144 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:27 crc kubenswrapper[4906]: I1123 06:51:27.356167 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:27 crc kubenswrapper[4906]: I1123 06:51:27.356191 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:27 crc kubenswrapper[4906]: E1123 06:51:27.358087 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:27 crc kubenswrapper[4906]: E1123 06:51:27.358964 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:27 crc kubenswrapper[4906]: E1123 06:51:27.359144 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:28 crc kubenswrapper[4906]: I1123 06:51:28.356005 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:28 crc kubenswrapper[4906]: E1123 06:51:28.356172 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:29 crc kubenswrapper[4906]: I1123 06:51:29.355827 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:29 crc kubenswrapper[4906]: E1123 06:51:29.355980 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:29 crc kubenswrapper[4906]: I1123 06:51:29.356087 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:29 crc kubenswrapper[4906]: I1123 06:51:29.356272 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:29 crc kubenswrapper[4906]: E1123 06:51:29.356414 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:29 crc kubenswrapper[4906]: E1123 06:51:29.356506 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:30 crc kubenswrapper[4906]: I1123 06:51:30.356030 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:30 crc kubenswrapper[4906]: E1123 06:51:30.357350 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:31 crc kubenswrapper[4906]: I1123 06:51:31.356471 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:31 crc kubenswrapper[4906]: I1123 06:51:31.356536 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:31 crc kubenswrapper[4906]: I1123 06:51:31.356492 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:31 crc kubenswrapper[4906]: E1123 06:51:31.356920 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:31 crc kubenswrapper[4906]: E1123 06:51:31.357091 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:31 crc kubenswrapper[4906]: E1123 06:51:31.357230 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:32 crc kubenswrapper[4906]: I1123 06:51:32.355749 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:32 crc kubenswrapper[4906]: E1123 06:51:32.356019 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:33 crc kubenswrapper[4906]: I1123 06:51:33.356046 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:33 crc kubenswrapper[4906]: I1123 06:51:33.356090 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:33 crc kubenswrapper[4906]: E1123 06:51:33.356274 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:33 crc kubenswrapper[4906]: I1123 06:51:33.356298 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:33 crc kubenswrapper[4906]: E1123 06:51:33.356428 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:33 crc kubenswrapper[4906]: E1123 06:51:33.356560 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:34 crc kubenswrapper[4906]: I1123 06:51:34.356415 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:34 crc kubenswrapper[4906]: E1123 06:51:34.356648 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:35 crc kubenswrapper[4906]: I1123 06:51:35.355547 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:35 crc kubenswrapper[4906]: I1123 06:51:35.355725 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:35 crc kubenswrapper[4906]: I1123 06:51:35.357758 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:35 crc kubenswrapper[4906]: E1123 06:51:35.357754 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:35 crc kubenswrapper[4906]: E1123 06:51:35.357911 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:35 crc kubenswrapper[4906]: E1123 06:51:35.358048 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:36 crc kubenswrapper[4906]: I1123 06:51:36.356127 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:36 crc kubenswrapper[4906]: E1123 06:51:36.356337 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:37 crc kubenswrapper[4906]: I1123 06:51:37.356193 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:37 crc kubenswrapper[4906]: E1123 06:51:37.356367 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:37 crc kubenswrapper[4906]: I1123 06:51:37.356491 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:37 crc kubenswrapper[4906]: I1123 06:51:37.356554 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:37 crc kubenswrapper[4906]: E1123 06:51:37.356751 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:37 crc kubenswrapper[4906]: E1123 06:51:37.356978 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:38 crc kubenswrapper[4906]: I1123 06:51:38.356567 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:38 crc kubenswrapper[4906]: E1123 06:51:38.356838 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:38 crc kubenswrapper[4906]: I1123 06:51:38.358280 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.094170 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/1.log" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.095029 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/0.log" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.095109 4906 generic.go:334] "Generic (PLEG): container finished" podID="4f75f165-ef56-40be-bfd4-1843de92b356" containerID="9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28" exitCode=1 Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.095223 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerDied","Data":"9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28"} Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.095330 4906 scope.go:117] "RemoveContainer" containerID="2a07d0d43f3f5173a125993ae18c4a12bbe82ca37fec4933aa5c061f93a39fd5" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.095929 4906 scope.go:117] "RemoveContainer" containerID="9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28" Nov 23 06:51:39 crc kubenswrapper[4906]: E1123 06:51:39.096247 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-lm9xt_openshift-multus(4f75f165-ef56-40be-bfd4-1843de92b356)\"" pod="openshift-multus/multus-lm9xt" podUID="4f75f165-ef56-40be-bfd4-1843de92b356" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.101125 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/3.log" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.105510 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerStarted","Data":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.106138 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.158937 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podStartSLOduration=95.158914819 podStartE2EDuration="1m35.158914819s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:51:39.156927157 +0000 UTC m=+114.670318470" watchObservedRunningTime="2025-11-23 06:51:39.158914819 +0000 UTC m=+114.672306132" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.356296 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.356436 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:39 crc kubenswrapper[4906]: E1123 06:51:39.356488 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.356549 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:39 crc kubenswrapper[4906]: E1123 06:51:39.356756 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:39 crc kubenswrapper[4906]: E1123 06:51:39.357105 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.473961 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8drn8"] Nov 23 06:51:39 crc kubenswrapper[4906]: I1123 06:51:39.474162 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:39 crc kubenswrapper[4906]: E1123 06:51:39.474322 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:40 crc kubenswrapper[4906]: I1123 06:51:40.111530 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/1.log" Nov 23 06:51:41 crc kubenswrapper[4906]: I1123 06:51:41.356044 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:41 crc kubenswrapper[4906]: I1123 06:51:41.356069 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:41 crc kubenswrapper[4906]: I1123 06:51:41.356115 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:41 crc kubenswrapper[4906]: I1123 06:51:41.356163 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:41 crc kubenswrapper[4906]: E1123 06:51:41.356806 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:41 crc kubenswrapper[4906]: E1123 06:51:41.356490 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:41 crc kubenswrapper[4906]: E1123 06:51:41.356958 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:41 crc kubenswrapper[4906]: E1123 06:51:41.357027 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:43 crc kubenswrapper[4906]: I1123 06:51:43.355536 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:43 crc kubenswrapper[4906]: I1123 06:51:43.355809 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:43 crc kubenswrapper[4906]: E1123 06:51:43.355940 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:43 crc kubenswrapper[4906]: I1123 06:51:43.356034 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:43 crc kubenswrapper[4906]: I1123 06:51:43.356046 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:43 crc kubenswrapper[4906]: E1123 06:51:43.356204 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:43 crc kubenswrapper[4906]: E1123 06:51:43.356342 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:43 crc kubenswrapper[4906]: E1123 06:51:43.356436 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:45 crc kubenswrapper[4906]: E1123 06:51:45.312412 4906 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 23 06:51:45 crc kubenswrapper[4906]: I1123 06:51:45.355644 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:45 crc kubenswrapper[4906]: I1123 06:51:45.355836 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:45 crc kubenswrapper[4906]: E1123 06:51:45.357377 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:45 crc kubenswrapper[4906]: I1123 06:51:45.357428 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:45 crc kubenswrapper[4906]: I1123 06:51:45.357465 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:45 crc kubenswrapper[4906]: E1123 06:51:45.357590 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:45 crc kubenswrapper[4906]: E1123 06:51:45.357832 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:45 crc kubenswrapper[4906]: E1123 06:51:45.358145 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:45 crc kubenswrapper[4906]: E1123 06:51:45.511432 4906 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 06:51:47 crc kubenswrapper[4906]: I1123 06:51:47.355835 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:47 crc kubenswrapper[4906]: I1123 06:51:47.355889 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:47 crc kubenswrapper[4906]: I1123 06:51:47.355836 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:47 crc kubenswrapper[4906]: E1123 06:51:47.355977 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:47 crc kubenswrapper[4906]: E1123 06:51:47.356125 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:47 crc kubenswrapper[4906]: I1123 06:51:47.356422 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:47 crc kubenswrapper[4906]: E1123 06:51:47.356407 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:47 crc kubenswrapper[4906]: E1123 06:51:47.356474 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:49 crc kubenswrapper[4906]: I1123 06:51:49.356263 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:49 crc kubenswrapper[4906]: I1123 06:51:49.356388 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:49 crc kubenswrapper[4906]: E1123 06:51:49.356450 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:49 crc kubenswrapper[4906]: E1123 06:51:49.356588 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:49 crc kubenswrapper[4906]: I1123 06:51:49.356741 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:49 crc kubenswrapper[4906]: E1123 06:51:49.356856 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:49 crc kubenswrapper[4906]: I1123 06:51:49.356921 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:49 crc kubenswrapper[4906]: E1123 06:51:49.357065 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:50 crc kubenswrapper[4906]: E1123 06:51:50.513558 4906 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 06:51:51 crc kubenswrapper[4906]: I1123 06:51:51.356303 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:51 crc kubenswrapper[4906]: I1123 06:51:51.356364 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:51 crc kubenswrapper[4906]: I1123 06:51:51.356319 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:51 crc kubenswrapper[4906]: I1123 06:51:51.356303 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:51 crc kubenswrapper[4906]: E1123 06:51:51.356496 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:51 crc kubenswrapper[4906]: E1123 06:51:51.356741 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:51 crc kubenswrapper[4906]: E1123 06:51:51.356830 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:51 crc kubenswrapper[4906]: E1123 06:51:51.356918 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:53 crc kubenswrapper[4906]: I1123 06:51:53.356335 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:53 crc kubenswrapper[4906]: I1123 06:51:53.356344 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:53 crc kubenswrapper[4906]: E1123 06:51:53.356941 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:53 crc kubenswrapper[4906]: I1123 06:51:53.356373 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:53 crc kubenswrapper[4906]: I1123 06:51:53.356344 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:53 crc kubenswrapper[4906]: E1123 06:51:53.357161 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:53 crc kubenswrapper[4906]: E1123 06:51:53.357300 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:53 crc kubenswrapper[4906]: E1123 06:51:53.357399 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:54 crc kubenswrapper[4906]: I1123 06:51:54.356160 4906 scope.go:117] "RemoveContainer" containerID="9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28" Nov 23 06:51:55 crc kubenswrapper[4906]: I1123 06:51:55.167077 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/1.log" Nov 23 06:51:55 crc kubenswrapper[4906]: I1123 06:51:55.167181 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerStarted","Data":"ce3c526adbece1506c82bf469d4fbe46fa7b6b3111c30f675503011841d55715"} Nov 23 06:51:55 crc kubenswrapper[4906]: I1123 06:51:55.355944 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:55 crc kubenswrapper[4906]: I1123 06:51:55.355944 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:55 crc kubenswrapper[4906]: I1123 06:51:55.355944 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:55 crc kubenswrapper[4906]: I1123 06:51:55.356076 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:55 crc kubenswrapper[4906]: E1123 06:51:55.358646 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:55 crc kubenswrapper[4906]: E1123 06:51:55.358803 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:55 crc kubenswrapper[4906]: E1123 06:51:55.359028 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:55 crc kubenswrapper[4906]: E1123 06:51:55.359173 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:55 crc kubenswrapper[4906]: E1123 06:51:55.514296 4906 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 06:51:57 crc kubenswrapper[4906]: I1123 06:51:57.355618 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:57 crc kubenswrapper[4906]: E1123 06:51:57.355884 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:51:57 crc kubenswrapper[4906]: I1123 06:51:57.356161 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:57 crc kubenswrapper[4906]: I1123 06:51:57.356226 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:57 crc kubenswrapper[4906]: I1123 06:51:57.356391 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:57 crc kubenswrapper[4906]: E1123 06:51:57.356373 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:57 crc kubenswrapper[4906]: E1123 06:51:57.356523 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:57 crc kubenswrapper[4906]: E1123 06:51:57.356731 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:59 crc kubenswrapper[4906]: I1123 06:51:59.356564 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:51:59 crc kubenswrapper[4906]: I1123 06:51:59.356613 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:51:59 crc kubenswrapper[4906]: I1123 06:51:59.356564 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:51:59 crc kubenswrapper[4906]: I1123 06:51:59.356715 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:51:59 crc kubenswrapper[4906]: E1123 06:51:59.356797 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:51:59 crc kubenswrapper[4906]: E1123 06:51:59.356928 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:51:59 crc kubenswrapper[4906]: E1123 06:51:59.357058 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8drn8" podUID="98bb1da0-e8ce-4b13-8a4d-c30423c5baa8" Nov 23 06:51:59 crc kubenswrapper[4906]: E1123 06:51:59.357422 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.355915 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.355968 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.356028 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.356120 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.359512 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.359646 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.359777 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.359773 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.360213 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 23 06:52:01 crc kubenswrapper[4906]: I1123 06:52:01.361225 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 23 06:52:03 crc kubenswrapper[4906]: I1123 06:52:03.963454 4906 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.009818 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lgph8"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.010396 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.019739 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.020641 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.020764 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.020783 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.022061 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.022510 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qdqtn"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.022541 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.023640 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.024893 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.025599 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.027208 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7ghjq"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.028300 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.028986 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.030143 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.033810 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-4qmrn"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.034853 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-4qmrn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.041591 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.042476 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044112 4906 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert": failed to list *v1.Secret: secrets "openshift-apiserver-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044174 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044371 4906 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config": failed to list *v1.ConfigMap: configmaps "openshift-apiserver-operator-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044406 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-apiserver-operator-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.044418 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044463 4906 reflector.go:561] object-"openshift-apiserver-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044492 4906 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044500 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044519 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.044574 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044667 4906 reflector.go:561] object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff": failed to list *v1.Secret: secrets "openshift-apiserver-sa-dockercfg-djjff" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044705 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-djjff\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-sa-dockercfg-djjff\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044720 4906 reflector.go:561] object-"openshift-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044758 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.044781 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044798 4906 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv": failed to list *v1.Secret: secrets "openshift-apiserver-operator-dockercfg-xtcjv" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044817 4906 reflector.go:561] object-"openshift-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044838 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044838 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-xtcjv\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-operator-dockercfg-xtcjv\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044806 4906 reflector.go:561] object-"openshift-oauth-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.044883 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.044895 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.044967 4906 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045008 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045037 4906 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045041 4906 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045057 4906 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045064 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045090 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045096 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045239 4906 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.045261 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045270 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045347 4906 reflector.go:561] object-"openshift-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045370 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045513 4906 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045525 4906 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045541 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045555 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045603 4906 reflector.go:561] object-"openshift-console"/"default-dockercfg-chnjx": failed to list *v1.Secret: secrets "default-dockercfg-chnjx" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045615 4906 reflector.go:561] object-"openshift-apiserver"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045625 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"default-dockercfg-chnjx\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-chnjx\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045635 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.045733 4906 reflector.go:561] object-"openshift-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.045760 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.045840 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.046018 4906 reflector.go:561] object-"openshift-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.046038 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.046179 4906 reflector.go:561] object-"openshift-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.046199 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.046416 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.047119 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.047116 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wp8fn"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.047882 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.048360 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-65gnv"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.049323 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.054771 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.055377 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.055735 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056110 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056290 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056452 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-config\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056468 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056515 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056599 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-client-ca\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056632 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qxmv\" (UniqueName: \"kubernetes.io/projected/aed5c517-2dbe-47af-9761-7979baf177b7-kube-api-access-8qxmv\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056647 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bfhxf"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.056666 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed5c517-2dbe-47af-9761-7979baf177b7-serving-cert\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.057329 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.057878 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.070168 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.071299 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.071599 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.084619 4906 reflector.go:561] object-"openshift-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.092587 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.092735 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.094868 4906 reflector.go:561] object-"openshift-console"/"console-config": failed to list *v1.ConfigMap: configmaps "console-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.094937 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"console-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.095576 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t"] Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.095655 4906 reflector.go:561] object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z": failed to list *v1.Secret: secrets "openshift-config-operator-dockercfg-7pc5z" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.095695 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-7pc5z\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-config-operator-dockercfg-7pc5z\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.095740 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.095894 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.095987 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096005 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096009 4906 reflector.go:561] object-"openshift-console"/"console-serving-cert": failed to list *v1.Secret: secrets "console-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096040 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096098 4906 reflector.go:561] object-"openshift-console"/"console-dockercfg-f62pw": failed to list *v1.Secret: secrets "console-dockercfg-f62pw" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096119 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096163 4906 reflector.go:561] object-"openshift-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096167 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096223 4906 reflector.go:561] object-"openshift-console"/"console-oauth-config": failed to list *v1.Secret: secrets "console-oauth-config" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096179 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096235 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-oauth-config\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-oauth-config\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096281 4906 reflector.go:561] object-"openshift-config-operator"/"config-operator-serving-cert": failed to list *v1.Secret: secrets "config-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096292 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"config-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"config-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096295 4906 reflector.go:561] object-"openshift-console"/"oauth-serving-cert": failed to list *v1.ConfigMap: configmaps "oauth-serving-cert" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096312 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"oauth-serving-cert\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"oauth-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096325 4906 reflector.go:561] object-"openshift-console"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: W1123 06:52:04.096349 4906 reflector.go:561] object-"openshift-console"/"service-ca": failed to list *v1.ConfigMap: configmaps "service-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096357 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096379 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"service-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"service-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096115 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-dockercfg-f62pw\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-dockercfg-f62pw\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: E1123 06:52:04.096351 4906 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096549 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096582 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096782 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096792 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.096899 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.097123 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.097291 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.097417 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.097560 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.097752 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.097950 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.099051 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.099817 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.104031 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.104028 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.104173 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.104804 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.105202 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.105303 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.105313 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.104180 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.106087 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.105775 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.106790 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ll4c9"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.107179 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.107533 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rwvsq"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.108137 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.109168 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-44r49"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.109434 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.115886 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.115982 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116427 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116521 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116596 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116667 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116766 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116880 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116999 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.117088 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.117001 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.117196 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.116717 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.117638 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.117923 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.118265 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.119394 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.121693 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.121802 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.121891 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.122080 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.122136 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.122353 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.122470 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.122576 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.122665 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.122674 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.125604 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.128546 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.130990 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.131479 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.132061 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.132315 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.132713 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.132977 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.135922 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jdr4r"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.138996 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.143849 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8hr2w"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.144725 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.146063 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.151028 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vhv4d"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.151574 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.156307 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qdqtn"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.156818 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.156608 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.178591 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lgph8"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.178850 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.179339 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.180720 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.181726 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.186380 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193046 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/11e47284-7e15-4254-93d2-8f2487db8c16-images\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193185 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/302e855d-0e92-4810-9ca5-f7c9bb4b549f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193255 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193310 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193336 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193354 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193402 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdfwm\" (UniqueName: \"kubernetes.io/projected/302e855d-0e92-4810-9ca5-f7c9bb4b549f-kube-api-access-wdfwm\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193439 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgd69\" (UniqueName: \"kubernetes.io/projected/b7452b46-2b11-4a4f-893e-22bd995151bc-kube-api-access-lgd69\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193485 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbbz5\" (UniqueName: \"kubernetes.io/projected/15d81e07-ddd3-436e-9848-14f986e9381f-kube-api-access-rbbz5\") pod \"downloads-7954f5f757-4qmrn\" (UID: \"15d81e07-ddd3-436e-9848-14f986e9381f\") " pod="openshift-console/downloads-7954f5f757-4qmrn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193511 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42xpz\" (UniqueName: \"kubernetes.io/projected/11e47284-7e15-4254-93d2-8f2487db8c16-kube-api-access-42xpz\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193570 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-config\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193595 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193641 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-metrics-certs\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193668 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-stats-auth\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193728 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2beee-ee56-4318-a0a8-8ed788d4fa51-config\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193796 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-client-ca\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193884 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f13d40d-942d-4e41-99cf-3630bbc65532-service-ca-bundle\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193917 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4619397-f891-4dc8-997e-cbb9c716c230-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193969 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-client\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.193991 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194045 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4619397-f891-4dc8-997e-cbb9c716c230-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194069 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80a2beee-ee56-4318-a0a8-8ed788d4fa51-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194135 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jxnt\" (UniqueName: \"kubernetes.io/projected/4f13d40d-942d-4e41-99cf-3630bbc65532-kube-api-access-8jxnt\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194162 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/302e855d-0e92-4810-9ca5-f7c9bb4b549f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194243 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqkcz\" (UniqueName: \"kubernetes.io/projected/71c94a3a-43d0-4513-9b90-352a44eba266-kube-api-access-zqkcz\") pod \"cluster-samples-operator-665b6dd947-qw8kr\" (UID: \"71c94a3a-43d0-4513-9b90-352a44eba266\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194297 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c94a3a-43d0-4513-9b90-352a44eba266-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qw8kr\" (UID: \"71c94a3a-43d0-4513-9b90-352a44eba266\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194333 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194383 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg2h5\" (UniqueName: \"kubernetes.io/projected/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-kube-api-access-sg2h5\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194411 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4619397-f891-4dc8-997e-cbb9c716c230-config\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194493 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11e47284-7e15-4254-93d2-8f2487db8c16-config\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194523 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-audit-policies\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194574 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/302e855d-0e92-4810-9ca5-f7c9bb4b549f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194591 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194607 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-default-certificate\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194645 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194661 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-client-ca\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194673 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194741 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qxmv\" (UniqueName: \"kubernetes.io/projected/aed5c517-2dbe-47af-9761-7979baf177b7-kube-api-access-8qxmv\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194778 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/11e47284-7e15-4254-93d2-8f2487db8c16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194822 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2beee-ee56-4318-a0a8-8ed788d4fa51-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194901 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-audit-dir\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194935 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194941 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed5c517-2dbe-47af-9761-7979baf177b7-serving-cert\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.194989 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.195369 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.195548 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-config\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.196586 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n49vg"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.197597 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.198124 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.198657 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.198865 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rft95"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.198965 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.199026 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.199672 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.200339 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed5c517-2dbe-47af-9761-7979baf177b7-serving-cert\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.200389 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ktpz"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.200991 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.201007 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.201541 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.204560 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bfhxf"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.207316 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.207833 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.208165 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rltxh"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.208544 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.209114 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.209256 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.210730 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-4qmrn"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.211591 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.214328 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.214652 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-65gnv"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.214695 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.214706 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.216772 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7ghjq"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.217054 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.219018 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.219972 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.220963 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rwvsq"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.221920 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.222860 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-5txhl"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.223953 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ll4c9"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.224092 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.224891 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jdr4r"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.225871 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wp8fn"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.226824 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zhqrw"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.227798 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.227878 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.228875 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n49vg"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.229886 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.231011 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.231904 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.232978 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.233638 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vhv4d"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.235356 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.236369 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rft95"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.237093 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.238602 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.239036 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.244357 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.247584 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rltxh"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.248864 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8hr2w"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.249922 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.253126 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.255928 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zhqrw"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.270950 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.272056 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ktpz"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.273476 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5txhl"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.275053 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-cv5cn"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.275653 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.275937 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-s4w7m"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.276333 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.277898 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s4w7m"] Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.292870 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.295855 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c94a3a-43d0-4513-9b90-352a44eba266-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qw8kr\" (UID: \"71c94a3a-43d0-4513-9b90-352a44eba266\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.295889 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4619397-f891-4dc8-997e-cbb9c716c230-config\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.295932 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.295950 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg2h5\" (UniqueName: \"kubernetes.io/projected/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-kube-api-access-sg2h5\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.295967 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11e47284-7e15-4254-93d2-8f2487db8c16-config\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.295999 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-audit-policies\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296016 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/302e855d-0e92-4810-9ca5-f7c9bb4b549f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296034 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296048 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-default-certificate\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296098 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296148 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/11e47284-7e15-4254-93d2-8f2487db8c16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296163 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2beee-ee56-4318-a0a8-8ed788d4fa51-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296183 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-audit-dir\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296200 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296238 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/11e47284-7e15-4254-93d2-8f2487db8c16-images\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296255 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/302e855d-0e92-4810-9ca5-f7c9bb4b549f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296318 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296342 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296357 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296391 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296410 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdfwm\" (UniqueName: \"kubernetes.io/projected/302e855d-0e92-4810-9ca5-f7c9bb4b549f-kube-api-access-wdfwm\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296428 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgd69\" (UniqueName: \"kubernetes.io/projected/b7452b46-2b11-4a4f-893e-22bd995151bc-kube-api-access-lgd69\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296484 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbbz5\" (UniqueName: \"kubernetes.io/projected/15d81e07-ddd3-436e-9848-14f986e9381f-kube-api-access-rbbz5\") pod \"downloads-7954f5f757-4qmrn\" (UID: \"15d81e07-ddd3-436e-9848-14f986e9381f\") " pod="openshift-console/downloads-7954f5f757-4qmrn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296504 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42xpz\" (UniqueName: \"kubernetes.io/projected/11e47284-7e15-4254-93d2-8f2487db8c16-kube-api-access-42xpz\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296524 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-metrics-certs\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296640 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-stats-auth\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296665 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2beee-ee56-4318-a0a8-8ed788d4fa51-config\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296717 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f13d40d-942d-4e41-99cf-3630bbc65532-service-ca-bundle\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296737 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4619397-f891-4dc8-997e-cbb9c716c230-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296782 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-client\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296799 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296821 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4619397-f891-4dc8-997e-cbb9c716c230-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296861 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80a2beee-ee56-4318-a0a8-8ed788d4fa51-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.296997 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jxnt\" (UniqueName: \"kubernetes.io/projected/4f13d40d-942d-4e41-99cf-3630bbc65532-kube-api-access-8jxnt\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.297022 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/302e855d-0e92-4810-9ca5-f7c9bb4b549f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.297066 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqkcz\" (UniqueName: \"kubernetes.io/projected/71c94a3a-43d0-4513-9b90-352a44eba266-kube-api-access-zqkcz\") pod \"cluster-samples-operator-665b6dd947-qw8kr\" (UID: \"71c94a3a-43d0-4513-9b90-352a44eba266\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.297953 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4619397-f891-4dc8-997e-cbb9c716c230-config\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.297976 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f13d40d-942d-4e41-99cf-3630bbc65532-service-ca-bundle\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.298096 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.298758 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-audit-dir\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.298884 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/11e47284-7e15-4254-93d2-8f2487db8c16-images\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.299196 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-audit-policies\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.299268 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11e47284-7e15-4254-93d2-8f2487db8c16-config\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.299339 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/302e855d-0e92-4810-9ca5-f7c9bb4b549f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.299514 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a2beee-ee56-4318-a0a8-8ed788d4fa51-config\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.300901 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a2beee-ee56-4318-a0a8-8ed788d4fa51-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.301199 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-default-certificate\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.301274 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-metrics-certs\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.301341 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/302e855d-0e92-4810-9ca5-f7c9bb4b549f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.302451 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4619397-f891-4dc8-997e-cbb9c716c230-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.304233 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c94a3a-43d0-4513-9b90-352a44eba266-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qw8kr\" (UID: \"71c94a3a-43d0-4513-9b90-352a44eba266\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.304741 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/11e47284-7e15-4254-93d2-8f2487db8c16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.312186 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.318826 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4f13d40d-942d-4e41-99cf-3630bbc65532-stats-auth\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.332544 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.352877 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.372944 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.392591 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.414016 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.432156 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.463532 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.472738 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.492909 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.512625 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.533012 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.552419 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.573103 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.592163 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.612562 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.631974 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.653742 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.673140 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.693152 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.712390 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.731989 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.758020 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.772936 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.793670 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.833881 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.834010 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.853031 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.873083 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.893035 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.914155 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.933088 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.961397 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.973024 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 23 06:52:04 crc kubenswrapper[4906]: I1123 06:52:04.993451 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.013385 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.032739 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.053327 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.074003 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.103045 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.114253 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.132541 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.153614 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.173250 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.191811 4906 request.go:700] Waited for 1.010681927s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-operator-dockercfg-2bh8d&limit=500&resourceVersion=0 Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.193616 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.214398 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.234016 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.252259 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.273150 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297356 4906 configmap.go:193] Couldn't get configMap openshift-console/service-ca: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297443 4906 configmap.go:193] Couldn't get configMap openshift-console/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297551 4906 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297492 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.797438407 +0000 UTC m=+141.310829750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca" (UniqueName: "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297577 4906 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297672 4906 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297754 4906 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297661 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.797629873 +0000 UTC m=+141.311021206 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297872 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-client podName:0a2fd331-4b06-4d1b-94f1-673ea64e9d92 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.797852619 +0000 UTC m=+141.311243952 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-client") pod "apiserver-7bbb656c7d-tc2st" (UID: "0a2fd331-4b06-4d1b-94f1-673ea64e9d92") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297897 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.797884379 +0000 UTC m=+141.311275712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297921 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config podName:0a2fd331-4b06-4d1b-94f1-673ea64e9d92 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.79790946 +0000 UTC m=+141.311300803 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config") pod "apiserver-7bbb656c7d-tc2st" (UID: "0a2fd331-4b06-4d1b-94f1-673ea64e9d92") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.297946 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert podName:0a2fd331-4b06-4d1b-94f1-673ea64e9d92 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.797934251 +0000 UTC m=+141.311325584 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert") pod "apiserver-7bbb656c7d-tc2st" (UID: "0a2fd331-4b06-4d1b-94f1-673ea64e9d92") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.298486 4906 configmap.go:193] Couldn't get configMap openshift-console/oauth-serving-cert: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.298605 4906 configmap.go:193] Couldn't get configMap openshift-console/console-config: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.298508 4906 secret.go:188] Couldn't get secret openshift-console/console-oauth-config: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.298748 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.798714421 +0000 UTC m=+141.312105764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-config" (UniqueName: "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.298560 4906 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.298787 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.798772863 +0000 UTC m=+141.312164206 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-oauth-config" (UniqueName: "kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.298826 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle podName:0a2fd331-4b06-4d1b-94f1-673ea64e9d92 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.798810814 +0000 UTC m=+141.312202157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle") pod "apiserver-7bbb656c7d-tc2st" (UID: "0a2fd331-4b06-4d1b-94f1-673ea64e9d92") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: E1123 06:52:05.299197 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:05.799170394 +0000 UTC m=+141.312561937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "oauth-serving-cert" (UniqueName: "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.313123 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.313965 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qxmv\" (UniqueName: \"kubernetes.io/projected/aed5c517-2dbe-47af-9761-7979baf177b7-kube-api-access-8qxmv\") pod \"controller-manager-879f6c89f-lgph8\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.332228 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.352854 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.373999 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.392889 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.413583 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.434119 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.454167 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.473730 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.492977 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.513654 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.533094 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.553899 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.554020 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.573215 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.595029 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.613996 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.634213 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.673213 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.677044 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.694519 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.713954 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.733006 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.755222 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.772131 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.786374 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lgph8"] Nov 23 06:52:05 crc kubenswrapper[4906]: W1123 06:52:05.796099 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaed5c517_2dbe_47af_9761_7979baf177b7.slice/crio-0d813a908bf9f1e7c3adfa517ef48898bfdf9c6b46dcadcaa8121d597878d519 WatchSource:0}: Error finding container 0d813a908bf9f1e7c3adfa517ef48898bfdf9c6b46dcadcaa8121d597878d519: Status 404 returned error can't find the container with id 0d813a908bf9f1e7c3adfa517ef48898bfdf9c6b46dcadcaa8121d597878d519 Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.806573 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.813346 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817108 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817173 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817208 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817236 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817323 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817382 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817410 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817455 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817634 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-client\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.817669 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.848414 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.855935 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.873656 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.893115 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.913300 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.932051 4906 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 23 06:52:05 crc kubenswrapper[4906]: I1123 06:52:05.952610 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.013363 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033037 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9efeaef-505f-4243-a2a2-e456fcc85218-node-pullsecrets\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033104 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-config\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033125 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-serving-cert\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033165 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/747bba30-ab93-469d-82c3-8c0e464cf969-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033175 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033200 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-config\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033219 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033262 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbbdk\" (UniqueName: \"kubernetes.io/projected/b8b1a645-3eb3-461a-880a-28ab9003a2f0-kube-api-access-zbbdk\") pod \"dns-operator-744455d44c-rwvsq\" (UID: \"b8b1a645-3eb3-461a-880a-28ab9003a2f0\") " pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033302 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033319 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-audit\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033333 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9efeaef-505f-4243-a2a2-e456fcc85218-audit-dir\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033363 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-certificates\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033384 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-etcd-client\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033453 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033479 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-client-ca\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033498 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5efaa1-d4d3-4971-9afb-1490732014d6-serving-cert\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033516 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b8b1a645-3eb3-461a-880a-28ab9003a2f0-metrics-tls\") pod \"dns-operator-744455d44c-rwvsq\" (UID: \"b8b1a645-3eb3-461a-880a-28ab9003a2f0\") " pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033580 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-config\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033624 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2ks4\" (UniqueName: \"kubernetes.io/projected/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-kube-api-access-l2ks4\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033646 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/fea909e5-6cef-4949-ad98-a37df3ebe87e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033664 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg5zf\" (UniqueName: \"kubernetes.io/projected/fea909e5-6cef-4949-ad98-a37df3ebe87e-kube-api-access-gg5zf\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033730 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-trusted-ca\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033747 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njk6q\" (UniqueName: \"kubernetes.io/projected/2d46f022-7d61-4f91-9834-b8e3670ca17f-kube-api-access-njk6q\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033808 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbqlw\" (UniqueName: \"kubernetes.io/projected/c9efeaef-505f-4243-a2a2-e456fcc85218-kube-api-access-dbqlw\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033852 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033886 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-serving-cert\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033900 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/0b5efaa1-d4d3-4971-9afb-1490732014d6-kube-api-access-tmxtz\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033915 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-encryption-config\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033934 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033970 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.033988 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-bound-sa-token\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034006 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-etcd-serving-ca\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034025 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d46f022-7d61-4f91-9834-b8e3670ca17f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034040 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d46f022-7d61-4f91-9834-b8e3670ca17f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034065 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-image-import-ca\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034111 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgr25\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-kube-api-access-rgr25\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034252 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-machine-approver-tls\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034280 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwzj2\" (UniqueName: \"kubernetes.io/projected/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-kube-api-access-gwzj2\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034338 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fea909e5-6cef-4949-ad98-a37df3ebe87e-serving-cert\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034649 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/747bba30-ab93-469d-82c3-8c0e464cf969-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034712 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxzgc\" (UniqueName: \"kubernetes.io/projected/747bba30-ab93-469d-82c3-8c0e464cf969-kube-api-access-bxzgc\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034776 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-tls\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034820 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-config\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034869 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.034929 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-trusted-ca\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.035061 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-auth-proxy-config\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.035212 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:06.535196407 +0000 UTC m=+142.048587730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.053046 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.072597 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.093588 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.113016 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.136699 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.136844 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137174 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137240 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be5a3686-7427-4990-b4ec-6944a80bab27-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137282 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137333 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/747bba30-ab93-469d-82c3-8c0e464cf969-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137381 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj2qd\" (UniqueName: \"kubernetes.io/projected/35486e5f-4cac-4539-8395-8855abb73cf0-kube-api-access-vj2qd\") pod \"multus-admission-controller-857f4d67dd-n49vg\" (UID: \"35486e5f-4cac-4539-8395-8855abb73cf0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137433 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-auth-proxy-config\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.137465 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:06.637440303 +0000 UTC m=+142.150831776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137521 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n955p\" (UniqueName: \"kubernetes.io/projected/be5a3686-7427-4990-b4ec-6944a80bab27-kube-api-access-n955p\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137610 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9efeaef-505f-4243-a2a2-e456fcc85218-node-pullsecrets\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137643 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2rhx\" (UniqueName: \"kubernetes.io/projected/87eb0e66-14a8-4c87-8235-986e5bccec2b-kube-api-access-x2rhx\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138117 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c9efeaef-505f-4243-a2a2-e456fcc85218-node-pullsecrets\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.137675 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-config\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138465 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-serving-cert\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138504 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b944399b-370e-4acd-9be9-fda8415b0d44-tmpfs\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138520 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-auth-proxy-config\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138536 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/803fc706-4690-447f-9aa8-bb512d9136f2-audit-dir\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138560 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f49e922-260c-4308-b2f9-e16268dbaf85-serving-cert\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138603 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-registration-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138768 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8016f37-c4f6-4619-8b05-8b32b2f6e443-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kmzlj\" (UID: \"b8016f37-c4f6-4619-8b05-8b32b2f6e443\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138820 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2xcz\" (UniqueName: \"kubernetes.io/projected/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-kube-api-access-r2xcz\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138884 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-config\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138922 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dc8a7161-8582-4a46-a920-a8c635957b0f-signing-key\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.138971 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139044 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbbdk\" (UniqueName: \"kubernetes.io/projected/b8b1a645-3eb3-461a-880a-28ab9003a2f0-kube-api-access-zbbdk\") pod \"dns-operator-744455d44c-rwvsq\" (UID: \"b8b1a645-3eb3-461a-880a-28ab9003a2f0\") " pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139077 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-audit-policies\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139122 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139143 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-audit\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139184 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-config-volume\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139203 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139223 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139239 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139289 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-metrics-tls\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139311 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-client-ca\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139329 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-service-ca\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139371 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139418 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139482 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6cmp\" (UniqueName: \"kubernetes.io/projected/380e25b3-93bc-4e4c-a82d-2adbb242ae15-kube-api-access-w6cmp\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139533 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-config\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139927 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87eb0e66-14a8-4c87-8235-986e5bccec2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140117 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-secret-volume\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140229 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-config\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.139979 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-config\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140291 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140170 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-config\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140149 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-client-ca\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140535 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/35486e5f-4cac-4539-8395-8855abb73cf0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n49vg\" (UID: \"35486e5f-4cac-4539-8395-8855abb73cf0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140591 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/747bba30-ab93-469d-82c3-8c0e464cf969-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140817 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-trusted-ca\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.140937 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgx2j\" (UniqueName: \"kubernetes.io/projected/c490ab53-76a1-473e-b007-2493b5dc2f2a-kube-api-access-pgx2j\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141047 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141178 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141311 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzd5g\" (UniqueName: \"kubernetes.io/projected/b8016f37-c4f6-4619-8b05-8b32b2f6e443-kube-api-access-xzd5g\") pod \"package-server-manager-789f6589d5-kmzlj\" (UID: \"b8016f37-c4f6-4619-8b05-8b32b2f6e443\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141417 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-serving-cert\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141524 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/0b5efaa1-d4d3-4971-9afb-1490732014d6-kube-api-access-tmxtz\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141634 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-encryption-config\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141787 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ngsc\" (UniqueName: \"kubernetes.io/projected/29b40410-6b1e-44ce-98a1-f5148a2ad676-kube-api-access-9ngsc\") pod \"control-plane-machine-set-operator-78cbb6b69f-bdrqj\" (UID: \"29b40410-6b1e-44ce-98a1-f5148a2ad676\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141922 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-trusted-ca\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.141930 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dc8a7161-8582-4a46-a920-a8c635957b0f-signing-cabundle\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142007 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b484e00f-69e1-45b2-8d6e-12dc17f4d267-node-bootstrap-token\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142025 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142043 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-config\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142069 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-etcd-serving-ca\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142137 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b944399b-370e-4acd-9be9-fda8415b0d44-webhook-cert\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142156 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/29b40410-6b1e-44ce-98a1-f5148a2ad676-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bdrqj\" (UID: \"29b40410-6b1e-44ce-98a1-f5148a2ad676\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142709 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-etcd-serving-ca\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142776 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-profile-collector-cert\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142810 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-482lt\" (UniqueName: \"kubernetes.io/projected/803fc706-4690-447f-9aa8-bb512d9136f2-kube-api-access-482lt\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142867 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-service-ca-bundle\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.142954 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-machine-approver-tls\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143217 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6255e00b-ec87-429b-bc92-1df3cb176a5e-cert\") pod \"ingress-canary-s4w7m\" (UID: \"6255e00b-ec87-429b-bc92-1df3cb176a5e\") " pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143253 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fea909e5-6cef-4949-ad98-a37df3ebe87e-serving-cert\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143275 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-srv-cert\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143384 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143440 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-config-volume\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143467 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-tls\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143910 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-config\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143944 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxzgc\" (UniqueName: \"kubernetes.io/projected/747bba30-ab93-469d-82c3-8c0e464cf969-kube-api-access-bxzgc\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143971 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smvhm\" (UniqueName: \"kubernetes.io/projected/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-kube-api-access-smvhm\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.143993 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22d56\" (UniqueName: \"kubernetes.io/projected/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-kube-api-access-22d56\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144017 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w25cw\" (UniqueName: \"kubernetes.io/projected/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-kube-api-access-w25cw\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-trusted-ca\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144177 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/87eb0e66-14a8-4c87-8235-986e5bccec2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144292 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/be5a3686-7427-4990-b4ec-6944a80bab27-images\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144343 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144387 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/747bba30-ab93-469d-82c3-8c0e464cf969-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144416 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv57q\" (UniqueName: \"kubernetes.io/projected/b484e00f-69e1-45b2-8d6e-12dc17f4d267-kube-api-access-jv57q\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144452 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-mountpoint-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144488 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-certificates\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144510 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9efeaef-505f-4243-a2a2-e456fcc85218-audit-dir\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144533 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-client\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144555 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144581 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-plugins-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144619 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144641 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144670 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-etcd-client\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144721 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f49e922-260c-4308-b2f9-e16268dbaf85-config\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144746 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144767 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-ca\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144791 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2m5l\" (UniqueName: \"kubernetes.io/projected/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-kube-api-access-z2m5l\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144825 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5efaa1-d4d3-4971-9afb-1490732014d6-serving-cert\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144851 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b8b1a645-3eb3-461a-880a-28ab9003a2f0-metrics-tls\") pod \"dns-operator-744455d44c-rwvsq\" (UID: \"b8b1a645-3eb3-461a-880a-28ab9003a2f0\") " pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144878 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvkdx\" (UniqueName: \"kubernetes.io/projected/b944399b-370e-4acd-9be9-fda8415b0d44-kube-api-access-dvkdx\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144908 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw9f4\" (UniqueName: \"kubernetes.io/projected/6255e00b-ec87-429b-bc92-1df3cb176a5e-kube-api-access-vw9f4\") pod \"ingress-canary-s4w7m\" (UID: \"6255e00b-ec87-429b-bc92-1df3cb176a5e\") " pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144934 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2ks4\" (UniqueName: \"kubernetes.io/projected/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-kube-api-access-l2ks4\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144957 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnnvc\" (UniqueName: \"kubernetes.io/projected/51d28f4b-a6a1-4698-9081-52eb141bd832-kube-api-access-fnnvc\") pod \"migrator-59844c95c7-vppvc\" (UID: \"51d28f4b-a6a1-4698-9081-52eb141bd832\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.144979 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145001 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8464\" (UniqueName: \"kubernetes.io/projected/8a5e4a09-9b16-435b-bc05-00d6d174ede3-kube-api-access-n8464\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145029 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/fea909e5-6cef-4949-ad98-a37df3ebe87e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145053 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg5zf\" (UniqueName: \"kubernetes.io/projected/fea909e5-6cef-4949-ad98-a37df3ebe87e-kube-api-access-gg5zf\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145064 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-trusted-ca\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145110 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78xr7\" (UniqueName: \"kubernetes.io/projected/3fd1d033-c7b0-44e5-88b8-55d7540079b8-kube-api-access-78xr7\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145411 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-config\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145443 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njk6q\" (UniqueName: \"kubernetes.io/projected/2d46f022-7d61-4f91-9834-b8e3670ca17f-kube-api-access-njk6q\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145463 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/380e25b3-93bc-4e4c-a82d-2adbb242ae15-serving-cert\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145484 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3fd1d033-c7b0-44e5-88b8-55d7540079b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.145515 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbqlw\" (UniqueName: \"kubernetes.io/projected/c9efeaef-505f-4243-a2a2-e456fcc85218-kube-api-access-dbqlw\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.146330 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.146418 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9efeaef-505f-4243-a2a2-e456fcc85218-audit-dir\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.146809 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-tls\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147138 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-serving-cert\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147154 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/fea909e5-6cef-4949-ad98-a37df3ebe87e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147253 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-csi-data-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147292 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww56b\" (UniqueName: \"kubernetes.io/projected/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-kube-api-access-ww56b\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147327 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-socket-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147383 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147409 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be5a3686-7427-4990-b4ec-6944a80bab27-proxy-tls\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147428 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147447 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-bound-sa-token\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147501 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147522 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsnfs\" (UniqueName: \"kubernetes.io/projected/dc8a7161-8582-4a46-a920-a8c635957b0f-kube-api-access-gsnfs\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147537 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147555 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d46f022-7d61-4f91-9834-b8e3670ca17f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147564 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-certificates\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147596 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d46f022-7d61-4f91-9834-b8e3670ca17f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147616 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/87eb0e66-14a8-4c87-8235-986e5bccec2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147633 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-image-import-ca\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147746 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-srv-cert\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147782 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgr25\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-kube-api-access-rgr25\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147800 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6crl\" (UniqueName: \"kubernetes.io/projected/8f49e922-260c-4308-b2f9-e16268dbaf85-kube-api-access-b6crl\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147816 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c490ab53-76a1-473e-b007-2493b5dc2f2a-serving-cert\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147855 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3fd1d033-c7b0-44e5-88b8-55d7540079b8-proxy-tls\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147875 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b484e00f-69e1-45b2-8d6e-12dc17f4d267-certs\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147895 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwzj2\" (UniqueName: \"kubernetes.io/projected/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-kube-api-access-gwzj2\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.147915 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b944399b-370e-4acd-9be9-fda8415b0d44-apiservice-cert\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.148098 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.148804 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-machine-approver-tls\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.149281 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5efaa1-d4d3-4971-9afb-1490732014d6-serving-cert\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.150456 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/747bba30-ab93-469d-82c3-8c0e464cf969-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.150789 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-etcd-client\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.152832 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.152980 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b8b1a645-3eb3-461a-880a-28ab9003a2f0-metrics-tls\") pod \"dns-operator-744455d44c-rwvsq\" (UID: \"b8b1a645-3eb3-461a-880a-28ab9003a2f0\") " pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.166576 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqkcz\" (UniqueName: \"kubernetes.io/projected/71c94a3a-43d0-4513-9b90-352a44eba266-kube-api-access-zqkcz\") pod \"cluster-samples-operator-665b6dd947-qw8kr\" (UID: \"71c94a3a-43d0-4513-9b90-352a44eba266\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.187874 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42xpz\" (UniqueName: \"kubernetes.io/projected/11e47284-7e15-4254-93d2-8f2487db8c16-kube-api-access-42xpz\") pod \"machine-api-operator-5694c8668f-qdqtn\" (UID: \"11e47284-7e15-4254-93d2-8f2487db8c16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.206417 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4619397-f891-4dc8-997e-cbb9c716c230-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cxdsb\" (UID: \"c4619397-f891-4dc8-997e-cbb9c716c230\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.211207 4906 request.go:700] Waited for 1.913901688s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.213775 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" event={"ID":"aed5c517-2dbe-47af-9761-7979baf177b7","Type":"ContainerStarted","Data":"59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c"} Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.213825 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" event={"ID":"aed5c517-2dbe-47af-9761-7979baf177b7","Type":"ContainerStarted","Data":"0d813a908bf9f1e7c3adfa517ef48898bfdf9c6b46dcadcaa8121d597878d519"} Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.214017 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.216889 4906 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lgph8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.216952 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" podUID="aed5c517-2dbe-47af-9761-7979baf177b7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248665 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dc8a7161-8582-4a46-a920-a8c635957b0f-signing-cabundle\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248737 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b484e00f-69e1-45b2-8d6e-12dc17f4d267-node-bootstrap-token\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248768 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248794 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-config\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248843 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b944399b-370e-4acd-9be9-fda8415b0d44-webhook-cert\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248866 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/29b40410-6b1e-44ce-98a1-f5148a2ad676-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bdrqj\" (UID: \"29b40410-6b1e-44ce-98a1-f5148a2ad676\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248888 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-profile-collector-cert\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248919 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-482lt\" (UniqueName: \"kubernetes.io/projected/803fc706-4690-447f-9aa8-bb512d9136f2-kube-api-access-482lt\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248944 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-service-ca-bundle\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.248974 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6255e00b-ec87-429b-bc92-1df3cb176a5e-cert\") pod \"ingress-canary-s4w7m\" (UID: \"6255e00b-ec87-429b-bc92-1df3cb176a5e\") " pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249002 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-srv-cert\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249026 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249051 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-config-volume\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249091 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smvhm\" (UniqueName: \"kubernetes.io/projected/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-kube-api-access-smvhm\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249115 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22d56\" (UniqueName: \"kubernetes.io/projected/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-kube-api-access-22d56\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249139 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w25cw\" (UniqueName: \"kubernetes.io/projected/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-kube-api-access-w25cw\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249162 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/87eb0e66-14a8-4c87-8235-986e5bccec2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249223 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/be5a3686-7427-4990-b4ec-6944a80bab27-images\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249264 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249296 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv57q\" (UniqueName: \"kubernetes.io/projected/b484e00f-69e1-45b2-8d6e-12dc17f4d267-kube-api-access-jv57q\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249321 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-mountpoint-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249354 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-client\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249445 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249464 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-plugins-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249481 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249496 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249515 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f49e922-260c-4308-b2f9-e16268dbaf85-config\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249531 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249546 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-ca\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249565 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2m5l\" (UniqueName: \"kubernetes.io/projected/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-kube-api-access-z2m5l\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249581 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvkdx\" (UniqueName: \"kubernetes.io/projected/b944399b-370e-4acd-9be9-fda8415b0d44-kube-api-access-dvkdx\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249601 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw9f4\" (UniqueName: \"kubernetes.io/projected/6255e00b-ec87-429b-bc92-1df3cb176a5e-kube-api-access-vw9f4\") pod \"ingress-canary-s4w7m\" (UID: \"6255e00b-ec87-429b-bc92-1df3cb176a5e\") " pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249623 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnnvc\" (UniqueName: \"kubernetes.io/projected/51d28f4b-a6a1-4698-9081-52eb141bd832-kube-api-access-fnnvc\") pod \"migrator-59844c95c7-vppvc\" (UID: \"51d28f4b-a6a1-4698-9081-52eb141bd832\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249655 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249689 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8464\" (UniqueName: \"kubernetes.io/projected/8a5e4a09-9b16-435b-bc05-00d6d174ede3-kube-api-access-n8464\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249715 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78xr7\" (UniqueName: \"kubernetes.io/projected/3fd1d033-c7b0-44e5-88b8-55d7540079b8-kube-api-access-78xr7\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249731 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-config\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249751 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/380e25b3-93bc-4e4c-a82d-2adbb242ae15-serving-cert\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249767 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3fd1d033-c7b0-44e5-88b8-55d7540079b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249793 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-csi-data-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249811 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww56b\" (UniqueName: \"kubernetes.io/projected/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-kube-api-access-ww56b\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249806 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dc8a7161-8582-4a46-a920-a8c635957b0f-signing-cabundle\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249827 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-socket-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249844 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be5a3686-7427-4990-b4ec-6944a80bab27-proxy-tls\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249867 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249887 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsnfs\" (UniqueName: \"kubernetes.io/projected/dc8a7161-8582-4a46-a920-a8c635957b0f-kube-api-access-gsnfs\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249904 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249929 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/87eb0e66-14a8-4c87-8235-986e5bccec2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249951 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-srv-cert\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249975 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6crl\" (UniqueName: \"kubernetes.io/projected/8f49e922-260c-4308-b2f9-e16268dbaf85-kube-api-access-b6crl\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.249991 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c490ab53-76a1-473e-b007-2493b5dc2f2a-serving-cert\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250008 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3fd1d033-c7b0-44e5-88b8-55d7540079b8-proxy-tls\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250023 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b484e00f-69e1-45b2-8d6e-12dc17f4d267-certs\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250044 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b944399b-370e-4acd-9be9-fda8415b0d44-apiservice-cert\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250097 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be5a3686-7427-4990-b4ec-6944a80bab27-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250216 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250246 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250268 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj2qd\" (UniqueName: \"kubernetes.io/projected/35486e5f-4cac-4539-8395-8855abb73cf0-kube-api-access-vj2qd\") pod \"multus-admission-controller-857f4d67dd-n49vg\" (UID: \"35486e5f-4cac-4539-8395-8855abb73cf0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250312 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n955p\" (UniqueName: \"kubernetes.io/projected/be5a3686-7427-4990-b4ec-6944a80bab27-kube-api-access-n955p\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250335 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2rhx\" (UniqueName: \"kubernetes.io/projected/87eb0e66-14a8-4c87-8235-986e5bccec2b-kube-api-access-x2rhx\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250356 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b944399b-370e-4acd-9be9-fda8415b0d44-tmpfs\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250373 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/803fc706-4690-447f-9aa8-bb512d9136f2-audit-dir\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250390 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f49e922-260c-4308-b2f9-e16268dbaf85-serving-cert\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250405 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-registration-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250425 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8016f37-c4f6-4619-8b05-8b32b2f6e443-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kmzlj\" (UID: \"b8016f37-c4f6-4619-8b05-8b32b2f6e443\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250441 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2xcz\" (UniqueName: \"kubernetes.io/projected/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-kube-api-access-r2xcz\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250467 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dc8a7161-8582-4a46-a920-a8c635957b0f-signing-key\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250486 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-audit-policies\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250545 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-config-volume\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250561 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250598 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250615 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250637 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-metrics-tls\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250653 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-service-ca\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250669 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250706 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6cmp\" (UniqueName: \"kubernetes.io/projected/380e25b3-93bc-4e4c-a82d-2adbb242ae15-kube-api-access-w6cmp\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250744 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87eb0e66-14a8-4c87-8235-986e5bccec2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250760 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-secret-volume\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250785 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/35486e5f-4cac-4539-8395-8855abb73cf0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n49vg\" (UID: \"35486e5f-4cac-4539-8395-8855abb73cf0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250804 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgx2j\" (UniqueName: \"kubernetes.io/projected/c490ab53-76a1-473e-b007-2493b5dc2f2a-kube-api-access-pgx2j\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250820 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250841 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzd5g\" (UniqueName: \"kubernetes.io/projected/b8016f37-c4f6-4619-8b05-8b32b2f6e443-kube-api-access-xzd5g\") pod \"package-server-manager-789f6589d5-kmzlj\" (UID: \"b8016f37-c4f6-4619-8b05-8b32b2f6e443\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250874 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ngsc\" (UniqueName: \"kubernetes.io/projected/29b40410-6b1e-44ce-98a1-f5148a2ad676-kube-api-access-9ngsc\") pod \"control-plane-machine-set-operator-78cbb6b69f-bdrqj\" (UID: \"29b40410-6b1e-44ce-98a1-f5148a2ad676\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.251070 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.251887 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-config-volume\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.252946 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/be5a3686-7427-4990-b4ec-6944a80bab27-images\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.253699 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-ca\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.253843 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b484e00f-69e1-45b2-8d6e-12dc17f4d267-node-bootstrap-token\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.255139 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-service-ca\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.255919 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b944399b-370e-4acd-9be9-fda8415b0d44-webhook-cert\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.256432 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.256851 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3fd1d033-c7b0-44e5-88b8-55d7540079b8-proxy-tls\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.256919 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-audit-policies\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.256985 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-secret-volume\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.257482 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-config-volume\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.257984 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6255e00b-ec87-429b-bc92-1df3cb176a5e-cert\") pod \"ingress-canary-s4w7m\" (UID: \"6255e00b-ec87-429b-bc92-1df3cb176a5e\") " pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.258037 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-srv-cert\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.258465 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-service-ca-bundle\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.259226 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/29b40410-6b1e-44ce-98a1-f5148a2ad676-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bdrqj\" (UID: \"29b40410-6b1e-44ce-98a1-f5148a2ad676\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.259571 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.260001 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-srv-cert\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.260285 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3fd1d033-c7b0-44e5-88b8-55d7540079b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.260301 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.260311 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b484e00f-69e1-45b2-8d6e-12dc17f4d267-certs\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.260734 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.261475 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.261799 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c490ab53-76a1-473e-b007-2493b5dc2f2a-serving-cert\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.262349 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-profile-collector-cert\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.262441 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f49e922-260c-4308-b2f9-e16268dbaf85-config\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.262883 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/803fc706-4690-447f-9aa8-bb512d9136f2-audit-dir\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.263197 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b944399b-370e-4acd-9be9-fda8415b0d44-tmpfs\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.263747 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be5a3686-7427-4990-b4ec-6944a80bab27-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.264400 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.264906 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dc8a7161-8582-4a46-a920-a8c635957b0f-signing-key\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.265259 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87eb0e66-14a8-4c87-8235-986e5bccec2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.265266 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/35486e5f-4cac-4539-8395-8855abb73cf0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n49vg\" (UID: \"35486e5f-4cac-4539-8395-8855abb73cf0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.250142 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/380e25b3-93bc-4e4c-a82d-2adbb242ae15-config\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.265449 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:06.765428086 +0000 UTC m=+142.278819589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.265769 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdfwm\" (UniqueName: \"kubernetes.io/projected/302e855d-0e92-4810-9ca5-f7c9bb4b549f-kube-api-access-wdfwm\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.265863 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b944399b-370e-4acd-9be9-fda8415b0d44-apiservice-cert\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.265938 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-csi-data-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.266540 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-config\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.266877 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.266953 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.267147 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-plugins-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.267348 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.267624 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c490ab53-76a1-473e-b007-2493b5dc2f2a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.267999 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-registration-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.268250 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-socket-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.268291 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/87eb0e66-14a8-4c87-8235-986e5bccec2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.268324 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8a5e4a09-9b16-435b-bc05-00d6d174ede3-mountpoint-dir\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.268935 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.269126 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.269482 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b8016f37-c4f6-4619-8b05-8b32b2f6e443-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kmzlj\" (UID: \"b8016f37-c4f6-4619-8b05-8b32b2f6e443\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.269543 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be5a3686-7427-4990-b4ec-6944a80bab27-proxy-tls\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.271452 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.272014 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.272526 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f49e922-260c-4308-b2f9-e16268dbaf85-serving-cert\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.272560 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.272736 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/380e25b3-93bc-4e4c-a82d-2adbb242ae15-etcd-client\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.272987 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.274229 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/380e25b3-93bc-4e4c-a82d-2adbb242ae15-serving-cert\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.274565 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-metrics-tls\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.282430 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgd69\" (UniqueName: \"kubernetes.io/projected/b7452b46-2b11-4a4f-893e-22bd995151bc-kube-api-access-lgd69\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.288172 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.292231 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbbz5\" (UniqueName: \"kubernetes.io/projected/15d81e07-ddd3-436e-9848-14f986e9381f-kube-api-access-rbbz5\") pod \"downloads-7954f5f757-4qmrn\" (UID: \"15d81e07-ddd3-436e-9848-14f986e9381f\") " pod="openshift-console/downloads-7954f5f757-4qmrn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.325517 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jxnt\" (UniqueName: \"kubernetes.io/projected/4f13d40d-942d-4e41-99cf-3630bbc65532-kube-api-access-8jxnt\") pod \"router-default-5444994796-44r49\" (UID: \"4f13d40d-942d-4e41-99cf-3630bbc65532\") " pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.342508 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80a2beee-ee56-4318-a0a8-8ed788d4fa51-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-72rm5\" (UID: \"80a2beee-ee56-4318-a0a8-8ed788d4fa51\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.346618 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/302e855d-0e92-4810-9ca5-f7c9bb4b549f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-sgdc6\" (UID: \"302e855d-0e92-4810-9ca5-f7c9bb4b549f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.351883 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.351992 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:06.851969314 +0000 UTC m=+142.365360627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.353162 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.353371 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.353785 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:06.853772072 +0000 UTC m=+142.367163385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.360267 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-audit\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.379215 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.380600 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d46f022-7d61-4f91-9834-b8e3670ca17f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.390352 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.397336 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.400028 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.430273 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.432454 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.434554 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.439148 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.441001 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.452856 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.455717 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.456477 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:06.956457469 +0000 UTC m=+142.469848772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.473139 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.481450 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.493048 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.499818 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr"] Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.513815 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.532444 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.554537 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.558645 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.561432 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.061408358 +0000 UTC m=+142.574799691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.574366 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.593176 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.598821 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.600651 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.614895 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.616895 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5"] Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.624994 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-serving-cert\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: W1123 06:52:06.626412 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80a2beee_ee56_4318_a0a8_8ed788d4fa51.slice/crio-a8743aa3f2aaa4bdf1a8e4e2f26dd1ed8888c8147635fd3231483dc3a5daf80f WatchSource:0}: Error finding container a8743aa3f2aaa4bdf1a8e4e2f26dd1ed8888c8147635fd3231483dc3a5daf80f: Status 404 returned error can't find the container with id a8743aa3f2aaa4bdf1a8e4e2f26dd1ed8888c8147635fd3231483dc3a5daf80f Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.633300 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.643104 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-etcd-client\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.654999 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.659081 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-4qmrn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.662271 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb"] Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.663789 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.664238 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.164196539 +0000 UTC m=+142.677587982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.664550 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.665000 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.164989259 +0000 UTC m=+142.678380572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: W1123 06:52:06.671438 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4619397_f891_4dc8_997e_cbb9c716c230.slice/crio-89d9f914e3c99dc3f902fd940807a068f9ce2e09d46228f8ac67a4b125f18b64 WatchSource:0}: Error finding container 89d9f914e3c99dc3f902fd940807a068f9ce2e09d46228f8ac67a4b125f18b64: Status 404 returned error can't find the container with id 89d9f914e3c99dc3f902fd940807a068f9ce2e09d46228f8ac67a4b125f18b64 Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.672258 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.688270 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9efeaef-505f-4243-a2a2-e456fcc85218-encryption-config\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.692968 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.704042 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qdqtn"] Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.714416 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.722606 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.735403 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.752985 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.757791 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg2h5\" (UniqueName: \"kubernetes.io/projected/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-kube-api-access-sg2h5\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.766172 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.766527 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.266495536 +0000 UTC m=+142.779886849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.767244 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.767627 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.267605395 +0000 UTC m=+142.780996708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.772279 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.793091 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.794860 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-config\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.812780 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.817869 4906 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.817979 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert podName:0a2fd331-4b06-4d1b-94f1-673ea64e9d92 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.817950797 +0000 UTC m=+143.331342100 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert") pod "apiserver-7bbb656c7d-tc2st" (UID: "0a2fd331-4b06-4d1b-94f1-673ea64e9d92") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818186 4906 secret.go:188] Couldn't get secret openshift-console/console-oauth-config: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818430 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.818385749 +0000 UTC m=+143.331777052 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "console-oauth-config" (UniqueName: "kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818477 4906 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818520 4906 configmap.go:193] Couldn't get configMap openshift-console/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.818548 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c9efeaef-505f-4243-a2a2-e456fcc85218-image-import-ca\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818524 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle podName:0a2fd331-4b06-4d1b-94f1-673ea64e9d92 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.818513443 +0000 UTC m=+143.331904966 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle") pod "apiserver-7bbb656c7d-tc2st" (UID: "0a2fd331-4b06-4d1b-94f1-673ea64e9d92") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818599 4906 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818622 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.818605325 +0000 UTC m=+143.331996618 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818563 4906 configmap.go:193] Couldn't get configMap openshift-console/oauth-serving-cert: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818643 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config podName:0a2fd331-4b06-4d1b-94f1-673ea64e9d92 nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.818635456 +0000 UTC m=+143.332026759 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config") pod "apiserver-7bbb656c7d-tc2st" (UID: "0a2fd331-4b06-4d1b-94f1-673ea64e9d92") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.818701 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert podName:b7452b46-2b11-4a4f-893e-22bd995151bc nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.818669337 +0000 UTC m=+143.332060640 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "oauth-serving-cert" (UniqueName: "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert") pod "console-f9d7485db-wp8fn" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.826042 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6"] Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.834281 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.852162 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.861245 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-4qmrn"] Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.868533 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.868792 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.368766763 +0000 UTC m=+142.882158066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.869617 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.869962 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.369950855 +0000 UTC m=+142.883342158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.878954 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.892577 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.913195 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.933658 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.948234 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fea909e5-6cef-4949-ad98-a37df3ebe87e-serving-cert\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.952599 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.970360 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.970519 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.470481074 +0000 UTC m=+142.983872387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.971548 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:06 crc kubenswrapper[4906]: E1123 06:52:06.971879 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.471871241 +0000 UTC m=+142.985262544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.972465 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.983510 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d46f022-7d61-4f91-9834-b8e3670ca17f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:06 crc kubenswrapper[4906]: I1123 06:52:06.992498 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.072188 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e98fdbb-8ec9-4bfc-9875-2cab77551f49-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c6gcz\" (UID: \"8e98fdbb-8ec9-4bfc-9875-2cab77551f49\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.073386 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.074526 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.574502038 +0000 UTC m=+143.087893351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.102239 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbbdk\" (UniqueName: \"kubernetes.io/projected/b8b1a645-3eb3-461a-880a-28ab9003a2f0-kube-api-access-zbbdk\") pod \"dns-operator-744455d44c-rwvsq\" (UID: \"b8b1a645-3eb3-461a-880a-28ab9003a2f0\") " pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.109283 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/0b5efaa1-d4d3-4971-9afb-1490732014d6-kube-api-access-tmxtz\") pod \"route-controller-manager-6576b87f9c-zbz6t\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.125472 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxzgc\" (UniqueName: \"kubernetes.io/projected/747bba30-ab93-469d-82c3-8c0e464cf969-kube-api-access-bxzgc\") pod \"openshift-controller-manager-operator-756b6f6bc6-mbwkx\" (UID: \"747bba30-ab93-469d-82c3-8c0e464cf969\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.145188 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2ks4\" (UniqueName: \"kubernetes.io/projected/a46c0856-8c0c-4d11-8bbf-f49b5a99af3e-kube-api-access-l2ks4\") pod \"console-operator-58897d9998-bfhxf\" (UID: \"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e\") " pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.157237 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg5zf\" (UniqueName: \"kubernetes.io/projected/fea909e5-6cef-4949-ad98-a37df3ebe87e-kube-api-access-gg5zf\") pod \"openshift-config-operator-7777fb866f-65gnv\" (UID: \"fea909e5-6cef-4949-ad98-a37df3ebe87e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.168115 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njk6q\" (UniqueName: \"kubernetes.io/projected/2d46f022-7d61-4f91-9834-b8e3670ca17f-kube-api-access-njk6q\") pod \"openshift-apiserver-operator-796bbdcf4f-g7tgv\" (UID: \"2d46f022-7d61-4f91-9834-b8e3670ca17f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.175899 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.176467 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.676449496 +0000 UTC m=+143.189840789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.186409 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbqlw\" (UniqueName: \"kubernetes.io/projected/c9efeaef-505f-4243-a2a2-e456fcc85218-kube-api-access-dbqlw\") pod \"apiserver-76f77b778f-7ghjq\" (UID: \"c9efeaef-505f-4243-a2a2-e456fcc85218\") " pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.207327 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-bound-sa-token\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.219389 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" event={"ID":"80a2beee-ee56-4318-a0a8-8ed788d4fa51","Type":"ContainerStarted","Data":"5b4c16f18dd20a0e3109af087899f83867661103fa3b46464c0a5cc3c33b5519"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.219634 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" event={"ID":"80a2beee-ee56-4318-a0a8-8ed788d4fa51","Type":"ContainerStarted","Data":"a8743aa3f2aaa4bdf1a8e4e2f26dd1ed8888c8147635fd3231483dc3a5daf80f"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.221214 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4qmrn" event={"ID":"15d81e07-ddd3-436e-9848-14f986e9381f","Type":"ContainerStarted","Data":"29b5e8dab73d9494f250bea2a39f2e7b6ad1e53e06373d262c05e23c6e265930"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.221259 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4qmrn" event={"ID":"15d81e07-ddd3-436e-9848-14f986e9381f","Type":"ContainerStarted","Data":"404859fe8496507cf7efa4b6eaf8c056ac6e27925262161cd795954fba8003c7"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.222026 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-4qmrn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.223795 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" event={"ID":"c4619397-f891-4dc8-997e-cbb9c716c230","Type":"ContainerStarted","Data":"0fcf06b85771958a22cbd3f4e480d520fc4286f1021520fb4225d139ae3a5f7d"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.224228 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" event={"ID":"c4619397-f891-4dc8-997e-cbb9c716c230","Type":"ContainerStarted","Data":"89d9f914e3c99dc3f902fd940807a068f9ce2e09d46228f8ac67a4b125f18b64"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.224535 4906 patch_prober.go:28] interesting pod/downloads-7954f5f757-4qmrn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.224581 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4qmrn" podUID="15d81e07-ddd3-436e-9848-14f986e9381f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.225447 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-44r49" event={"ID":"4f13d40d-942d-4e41-99cf-3630bbc65532","Type":"ContainerStarted","Data":"b6751787e76a15b49db4e0064fc37c545249a8e61b49df8cb998bf483fc62ab8"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.225574 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-44r49" event={"ID":"4f13d40d-942d-4e41-99cf-3630bbc65532","Type":"ContainerStarted","Data":"4f1e8c5aa5b7ec89bd6e42f7c6fb3aa51e2fa914496ac2294900e80be1e2b144"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.226929 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" event={"ID":"302e855d-0e92-4810-9ca5-f7c9bb4b549f","Type":"ContainerStarted","Data":"7319cc87e282a3ff3c776b4f5264767806ea5e7ee8e87ec6ded31653ad73c52f"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.226956 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" event={"ID":"302e855d-0e92-4810-9ca5-f7c9bb4b549f","Type":"ContainerStarted","Data":"dfce10f9d9ac10b4a6bb70f1c3163fb1f426a14d98904eebaec047ac378a7b77"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.228949 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" event={"ID":"71c94a3a-43d0-4513-9b90-352a44eba266","Type":"ContainerStarted","Data":"5862f72d1b62aed75ed44ae420514a4c435fbf91502f0ff0765686366a0a155b"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.229029 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" event={"ID":"71c94a3a-43d0-4513-9b90-352a44eba266","Type":"ContainerStarted","Data":"bc17fecf2d52916c70a358c107758018ddb190c44713b8df2f274759f9bbc06b"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.229044 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" event={"ID":"71c94a3a-43d0-4513-9b90-352a44eba266","Type":"ContainerStarted","Data":"ec48066e8c8b383373100d2692b694770c6937f578fe94487b8eff8e11121485"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.229846 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgr25\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-kube-api-access-rgr25\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.230961 4906 request.go:700] Waited for 1.082900822s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.231036 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" event={"ID":"11e47284-7e15-4254-93d2-8f2487db8c16","Type":"ContainerStarted","Data":"319704155cc4fb5ff6aa84d25d1a433b95561d5f0809f06fc1102916c40cef7e"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.231068 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" event={"ID":"11e47284-7e15-4254-93d2-8f2487db8c16","Type":"ContainerStarted","Data":"f5a05da8332160f6e78948674af4b20a668b22e6de733adf0efdaf784e1c188e"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.231081 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" event={"ID":"11e47284-7e15-4254-93d2-8f2487db8c16","Type":"ContainerStarted","Data":"7dc21eb3fd1e703bb31c2fa70f489d80d1fe14aa71e590389b8b819bda7981c1"} Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.237695 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.238912 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.251236 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.252239 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwzj2\" (UniqueName: \"kubernetes.io/projected/dee9c5bf-a6e8-46cf-8253-e1a52effd5dc-kube-api-access-gwzj2\") pod \"machine-approver-56656f9798-mjbrq\" (UID: \"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.263630 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.267761 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-482lt\" (UniqueName: \"kubernetes.io/projected/803fc706-4690-447f-9aa8-bb512d9136f2-kube-api-access-482lt\") pod \"oauth-openshift-558db77b4-8hr2w\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.269497 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.277634 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.277862 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.777828599 +0000 UTC m=+143.291219912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.278141 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.278583 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.778564868 +0000 UTC m=+143.291956171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.280355 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.290651 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ngsc\" (UniqueName: \"kubernetes.io/projected/29b40410-6b1e-44ce-98a1-f5148a2ad676-kube-api-access-9ngsc\") pod \"control-plane-machine-set-operator-78cbb6b69f-bdrqj\" (UID: \"29b40410-6b1e-44ce-98a1-f5148a2ad676\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.309060 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6crl\" (UniqueName: \"kubernetes.io/projected/8f49e922-260c-4308-b2f9-e16268dbaf85-kube-api-access-b6crl\") pod \"service-ca-operator-777779d784-rft95\" (UID: \"8f49e922-260c-4308-b2f9-e16268dbaf85\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.323385 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.342019 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.361804 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.362158 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smvhm\" (UniqueName: \"kubernetes.io/projected/a3ef32c5-fbdd-4689-bbd2-1daae2872cb6-kube-api-access-smvhm\") pod \"catalog-operator-68c6474976-h4mkr\" (UID: \"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.362286 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2m5l\" (UniqueName: \"kubernetes.io/projected/c6b6fbeb-f958-4638-a1ee-bba42048f6ff-kube-api-access-z2m5l\") pod \"kube-storage-version-migrator-operator-b67b599dd-c62nz\" (UID: \"c6b6fbeb-f958-4638-a1ee-bba42048f6ff\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.383246 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.383425 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.883366843 +0000 UTC m=+143.396758146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.383879 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.385121 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.885105989 +0000 UTC m=+143.398497492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.397976 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22d56\" (UniqueName: \"kubernetes.io/projected/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-kube-api-access-22d56\") pod \"collect-profiles-29398005-pzgzx\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.402962 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.418732 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.422356 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.434097 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.441268 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.465269 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.465291 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6cmp\" (UniqueName: \"kubernetes.io/projected/380e25b3-93bc-4e4c-a82d-2adbb242ae15-kube-api-access-w6cmp\") pod \"etcd-operator-b45778765-jdr4r\" (UID: \"380e25b3-93bc-4e4c-a82d-2adbb242ae15\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.465752 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/87eb0e66-14a8-4c87-8235-986e5bccec2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.470468 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2xcz\" (UniqueName: \"kubernetes.io/projected/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-kube-api-access-r2xcz\") pod \"marketplace-operator-79b997595-8ktpz\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.472162 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w25cw\" (UniqueName: \"kubernetes.io/projected/4ad26aa7-4af7-4f02-87c6-6a6ce70bd381-kube-api-access-w25cw\") pod \"dns-default-5txhl\" (UID: \"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381\") " pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.474398 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.483792 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.486358 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.487050 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:07.987029127 +0000 UTC m=+143.500420430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.493184 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgx2j\" (UniqueName: \"kubernetes.io/projected/c490ab53-76a1-473e-b007-2493b5dc2f2a-kube-api-access-pgx2j\") pod \"authentication-operator-69f744f599-vhv4d\" (UID: \"c490ab53-76a1-473e-b007-2493b5dc2f2a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.499101 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvkdx\" (UniqueName: \"kubernetes.io/projected/b944399b-370e-4acd-9be9-fda8415b0d44-kube-api-access-dvkdx\") pod \"packageserver-d55dfcdfc-8hwdl\" (UID: \"b944399b-370e-4acd-9be9-fda8415b0d44\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.509227 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.518078 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.522696 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw9f4\" (UniqueName: \"kubernetes.io/projected/6255e00b-ec87-429b-bc92-1df3cb176a5e-kube-api-access-vw9f4\") pod \"ingress-canary-s4w7m\" (UID: \"6255e00b-ec87-429b-bc92-1df3cb176a5e\") " pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.529890 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:07 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:07 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:07 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.529953 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.530859 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.564570 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnnvc\" (UniqueName: \"kubernetes.io/projected/51d28f4b-a6a1-4698-9081-52eb141bd832-kube-api-access-fnnvc\") pod \"migrator-59844c95c7-vppvc\" (UID: \"51d28f4b-a6a1-4698-9081-52eb141bd832\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.567469 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzd5g\" (UniqueName: \"kubernetes.io/projected/b8016f37-c4f6-4619-8b05-8b32b2f6e443-kube-api-access-xzd5g\") pod \"package-server-manager-789f6589d5-kmzlj\" (UID: \"b8016f37-c4f6-4619-8b05-8b32b2f6e443\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.569129 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s4w7m" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.588408 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.588925 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.088911803 +0000 UTC m=+143.602303106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.611225 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n955p\" (UniqueName: \"kubernetes.io/projected/be5a3686-7427-4990-b4ec-6944a80bab27-kube-api-access-n955p\") pod \"machine-config-operator-74547568cd-h48nj\" (UID: \"be5a3686-7427-4990-b4ec-6944a80bab27\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.611555 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78xr7\" (UniqueName: \"kubernetes.io/projected/3fd1d033-c7b0-44e5-88b8-55d7540079b8-kube-api-access-78xr7\") pod \"machine-config-controller-84d6567774-tgwrc\" (UID: \"3fd1d033-c7b0-44e5-88b8-55d7540079b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.630778 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8464\" (UniqueName: \"kubernetes.io/projected/8a5e4a09-9b16-435b-bc05-00d6d174ede3-kube-api-access-n8464\") pod \"csi-hostpathplugin-zhqrw\" (UID: \"8a5e4a09-9b16-435b-bc05-00d6d174ede3\") " pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.638292 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2rhx\" (UniqueName: \"kubernetes.io/projected/87eb0e66-14a8-4c87-8235-986e5bccec2b-kube-api-access-x2rhx\") pod \"ingress-operator-5b745b69d9-cfqk7\" (UID: \"87eb0e66-14a8-4c87-8235-986e5bccec2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.652566 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.656002 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.659980 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww56b\" (UniqueName: \"kubernetes.io/projected/2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d-kube-api-access-ww56b\") pod \"olm-operator-6b444d44fb-tkds2\" (UID: \"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.682306 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.683093 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.684215 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj2qd\" (UniqueName: \"kubernetes.io/projected/35486e5f-4cac-4539-8395-8855abb73cf0-kube-api-access-vj2qd\") pod \"multus-admission-controller-857f4d67dd-n49vg\" (UID: \"35486e5f-4cac-4539-8395-8855abb73cf0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.685467 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-65gnv"] Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.698019 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.700765 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.701054 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.201019412 +0000 UTC m=+143.714410725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.701286 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.701595 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.201583767 +0000 UTC m=+143.714975220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.712139 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.732840 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv57q\" (UniqueName: \"kubernetes.io/projected/b484e00f-69e1-45b2-8d6e-12dc17f4d267-kube-api-access-jv57q\") pod \"machine-config-server-cv5cn\" (UID: \"b484e00f-69e1-45b2-8d6e-12dc17f4d267\") " pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.744598 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.780461 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.785115 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.801533 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsnfs\" (UniqueName: \"kubernetes.io/projected/dc8a7161-8582-4a46-a920-a8c635957b0f-kube-api-access-gsnfs\") pod \"service-ca-9c57cc56f-rltxh\" (UID: \"dc8a7161-8582-4a46-a920-a8c635957b0f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.801786 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.802103 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.302086817 +0000 UTC m=+143.815478120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.858867 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.864565 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-cv5cn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.907607 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.907704 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.907764 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.907791 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.907839 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.907861 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.907904 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.909138 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.910320 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.913255 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-encryption-config\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.913627 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2fd331-4b06-4d1b-94f1-673ea64e9d92-serving-cert\") pod \"apiserver-7bbb656c7d-tc2st\" (UID: \"0a2fd331-4b06-4d1b-94f1-673ea64e9d92\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.913923 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.914668 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle\") pod \"console-f9d7485db-wp8fn\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:07 crc kubenswrapper[4906]: E1123 06:52:07.915857 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.415843771 +0000 UTC m=+143.929235074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:07 crc kubenswrapper[4906]: I1123 06:52:07.962723 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx"] Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.008934 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.009224 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.509182039 +0000 UTC m=+144.022573502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.065351 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.097434 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.112764 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.113956 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.613942442 +0000 UTC m=+144.127333735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.121102 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:08 crc kubenswrapper[4906]: W1123 06:52:08.144497 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod747bba30_ab93_469d_82c3_8c0e464cf969.slice/crio-061760fd04848474a21827a048dbd9cebf27b397c7d2db68e021caf05c41c6cd WatchSource:0}: Error finding container 061760fd04848474a21827a048dbd9cebf27b397c7d2db68e021caf05c41c6cd: Status 404 returned error can't find the container with id 061760fd04848474a21827a048dbd9cebf27b397c7d2db68e021caf05c41c6cd Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.214482 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.214794 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.71477765 +0000 UTC m=+144.228168953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.237428 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-4qmrn" podStartSLOduration=124.237411274 podStartE2EDuration="2m4.237411274s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:08.234574208 +0000 UTC m=+143.747965511" watchObservedRunningTime="2025-11-23 06:52:08.237411274 +0000 UTC m=+143.750802577" Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.254618 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" event={"ID":"fea909e5-6cef-4949-ad98-a37df3ebe87e","Type":"ContainerStarted","Data":"74698fc3a0427b7290a954f7a1a929ec30cd648c1fd4c8a2039f7db5f46c4692"} Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.258476 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" event={"ID":"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc","Type":"ContainerStarted","Data":"740d36620e2524040774bdb9aa48e21a840c3162cdcb0a9ada8c4184e3e5b718"} Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.258513 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" event={"ID":"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc","Type":"ContainerStarted","Data":"8ddfd46553fad9ebe5f79a55593a3749474a86d641b78f28af02845cbd56ba1d"} Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.269002 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" event={"ID":"747bba30-ab93-469d-82c3-8c0e464cf969","Type":"ContainerStarted","Data":"061760fd04848474a21827a048dbd9cebf27b397c7d2db68e021caf05c41c6cd"} Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.271479 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cv5cn" event={"ID":"b484e00f-69e1-45b2-8d6e-12dc17f4d267","Type":"ContainerStarted","Data":"1356b30a7f24ce89851542adb4e05ae1cab9419fd0a92431922eb8a4d1084eea"} Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.274233 4906 patch_prober.go:28] interesting pod/downloads-7954f5f757-4qmrn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.274271 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4qmrn" podUID="15d81e07-ddd3-436e-9848-14f986e9381f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.317201 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.317548 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.81753604 +0000 UTC m=+144.330927343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.425068 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.425717 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.925660913 +0000 UTC m=+144.439052216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.426080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.444982 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:08.944955777 +0000 UTC m=+144.458347080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.462153 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bfhxf"] Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.529835 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.530457 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.030424316 +0000 UTC m=+144.543815619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.626547 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxdsb" podStartSLOduration=124.626526549 podStartE2EDuration="2m4.626526549s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:08.624729531 +0000 UTC m=+144.138120834" watchObservedRunningTime="2025-11-23 06:52:08.626526549 +0000 UTC m=+144.139917852" Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.631568 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.631937 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.131923223 +0000 UTC m=+144.645314526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.732479 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.732839 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.232818973 +0000 UTC m=+144.746210276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.816630 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" podStartSLOduration=124.816610927 podStartE2EDuration="2m4.816610927s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:08.775161561 +0000 UTC m=+144.288552864" watchObservedRunningTime="2025-11-23 06:52:08.816610927 +0000 UTC m=+144.330002230" Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.834880 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.835350 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.335333386 +0000 UTC m=+144.848724689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.853710 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-sgdc6" podStartSLOduration=124.85367200499999 podStartE2EDuration="2m4.853672005s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:08.816145645 +0000 UTC m=+144.329536948" watchObservedRunningTime="2025-11-23 06:52:08.853672005 +0000 UTC m=+144.367063308" Nov 23 06:52:08 crc kubenswrapper[4906]: I1123 06:52:08.936538 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:08 crc kubenswrapper[4906]: E1123 06:52:08.937330 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.437312605 +0000 UTC m=+144.950703908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:08 crc kubenswrapper[4906]: W1123 06:52:08.973046 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda46c0856_8c0c_4d11_8bbf_f49b5a99af3e.slice/crio-c9138587fb87963266962749ac3e91c37234d0e5a75c6f91d4dd7a775718174f WatchSource:0}: Error finding container c9138587fb87963266962749ac3e91c37234d0e5a75c6f91d4dd7a775718174f: Status 404 returned error can't find the container with id c9138587fb87963266962749ac3e91c37234d0e5a75c6f91d4dd7a775718174f Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.040195 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.040649 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.54063124 +0000 UTC m=+145.054022543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.049428 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qdqtn" podStartSLOduration=125.049407154 podStartE2EDuration="2m5.049407154s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.047791021 +0000 UTC m=+144.561182334" watchObservedRunningTime="2025-11-23 06:52:09.049407154 +0000 UTC m=+144.562798447" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.094272 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-72rm5" podStartSLOduration=125.094253169 podStartE2EDuration="2m5.094253169s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.09315084 +0000 UTC m=+144.606542143" watchObservedRunningTime="2025-11-23 06:52:09.094253169 +0000 UTC m=+144.607644472" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.142177 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.142599 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.642580398 +0000 UTC m=+145.155971701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.244358 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.244693 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.74466646 +0000 UTC m=+145.258057763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.275636 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" event={"ID":"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e","Type":"ContainerStarted","Data":"b4c6836c9b8ff0d83d832cb7743c776c4f1befb4849066b1d52dedb3f0bc996d"} Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.275696 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" event={"ID":"a46c0856-8c0c-4d11-8bbf-f49b5a99af3e","Type":"ContainerStarted","Data":"c9138587fb87963266962749ac3e91c37234d0e5a75c6f91d4dd7a775718174f"} Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.276408 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.277857 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" event={"ID":"dee9c5bf-a6e8-46cf-8253-e1a52effd5dc","Type":"ContainerStarted","Data":"4c406f3c405f7d805f21eb4920375801f35b154ff74cb67e0e7ae78d442014cd"} Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.280527 4906 patch_prober.go:28] interesting pod/console-operator-58897d9998-bfhxf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.280607 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" podUID="a46c0856-8c0c-4d11-8bbf-f49b5a99af3e" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.281481 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" event={"ID":"747bba30-ab93-469d-82c3-8c0e464cf969","Type":"ContainerStarted","Data":"cc02608f1aa87468010a0511bb777adf2c647359226b9d7f1418f436095db002"} Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.287494 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-cv5cn" event={"ID":"b484e00f-69e1-45b2-8d6e-12dc17f4d267","Type":"ContainerStarted","Data":"63ae95fa72788d15ad8c52f6f870b4c2acf1e74c85fe0a7f1343e10d7ab85932"} Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.289813 4906 generic.go:334] "Generic (PLEG): container finished" podID="fea909e5-6cef-4949-ad98-a37df3ebe87e" containerID="8212e6a207d8f6a245b168c5210675c8e37f04ce40a705687a148d5abda73868" exitCode=0 Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.290717 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" event={"ID":"fea909e5-6cef-4949-ad98-a37df3ebe87e","Type":"ContainerDied","Data":"8212e6a207d8f6a245b168c5210675c8e37f04ce40a705687a148d5abda73868"} Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.295759 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-44r49" podStartSLOduration=125.295739482 podStartE2EDuration="2m5.295739482s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.253200348 +0000 UTC m=+144.766591651" watchObservedRunningTime="2025-11-23 06:52:09.295739482 +0000 UTC m=+144.809130785" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.347588 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.347897 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.847880871 +0000 UTC m=+145.361272174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.348584 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.350744 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.850722657 +0000 UTC m=+145.364113960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.398667 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:09 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:09 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:09 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.398759 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.425748 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qw8kr" podStartSLOduration=125.425720167 podStartE2EDuration="2m5.425720167s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.408155729 +0000 UTC m=+144.921547032" watchObservedRunningTime="2025-11-23 06:52:09.425720167 +0000 UTC m=+144.939111470" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.460432 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.461789 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:09.961767339 +0000 UTC m=+145.475158642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.503057 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:09 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:09 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:09 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.503108 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.558479 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mbwkx" podStartSLOduration=125.558453176 podStartE2EDuration="2m5.558453176s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.548005267 +0000 UTC m=+145.061396570" watchObservedRunningTime="2025-11-23 06:52:09.558453176 +0000 UTC m=+145.071844479" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.566602 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.567059 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.067045275 +0000 UTC m=+145.580436578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.579099 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mjbrq" podStartSLOduration=125.579078546 podStartE2EDuration="2m5.579078546s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.578721696 +0000 UTC m=+145.092112999" watchObservedRunningTime="2025-11-23 06:52:09.579078546 +0000 UTC m=+145.092469849" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.583411 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj"] Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.635390 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv"] Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.647240 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz"] Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.668957 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" podStartSLOduration=125.668938342 podStartE2EDuration="2m5.668938342s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.667982427 +0000 UTC m=+145.181373730" watchObservedRunningTime="2025-11-23 06:52:09.668938342 +0000 UTC m=+145.182329645" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.669573 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.669672 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.169638461 +0000 UTC m=+145.683029754 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.669961 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.670415 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.170402471 +0000 UTC m=+145.683793764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.717074 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-cv5cn" podStartSLOduration=5.717038034 podStartE2EDuration="5.717038034s" podCreationTimestamp="2025-11-23 06:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:09.688921995 +0000 UTC m=+145.202313298" watchObservedRunningTime="2025-11-23 06:52:09.717038034 +0000 UTC m=+145.230429337" Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.771859 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.773107 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.272643757 +0000 UTC m=+145.786035060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.874911 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.875229 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.375217202 +0000 UTC m=+145.888608495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.928796 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rft95"] Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.930690 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t"] Nov 23 06:52:09 crc kubenswrapper[4906]: W1123 06:52:09.954727 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f49e922_260c_4308_b2f9_e16268dbaf85.slice/crio-56ffcf792074425cd2810839aa56db1ac3cf20e1e4f3984f8a6060bf2c616535 WatchSource:0}: Error finding container 56ffcf792074425cd2810839aa56db1ac3cf20e1e4f3984f8a6060bf2c616535: Status 404 returned error can't find the container with id 56ffcf792074425cd2810839aa56db1ac3cf20e1e4f3984f8a6060bf2c616535 Nov 23 06:52:09 crc kubenswrapper[4906]: W1123 06:52:09.956323 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b5efaa1_d4d3_4971_9afb_1490732014d6.slice/crio-1be19d1b1c054cf6ad8321ccb6efa014c39d4d959a24ec1496f775dd7ca6affb WatchSource:0}: Error finding container 1be19d1b1c054cf6ad8321ccb6efa014c39d4d959a24ec1496f775dd7ca6affb: Status 404 returned error can't find the container with id 1be19d1b1c054cf6ad8321ccb6efa014c39d4d959a24ec1496f775dd7ca6affb Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.962232 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj"] Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.967983 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jdr4r"] Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.982480 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:09 crc kubenswrapper[4906]: E1123 06:52:09.982951 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.482930594 +0000 UTC m=+145.996321897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.983449 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz"] Nov 23 06:52:09 crc kubenswrapper[4906]: I1123 06:52:09.986888 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.007722 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7ghjq"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.056860 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.069641 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.084609 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.085262 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.585246642 +0000 UTC m=+146.098637945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.148483 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wp8fn"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.185930 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.186333 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.686314116 +0000 UTC m=+146.199705429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.186967 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ktpz"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.192019 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rwvsq"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.194895 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.245222 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.256418 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.263787 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vhv4d"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.266368 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5txhl"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.271950 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.273457 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.275915 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8hr2w"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.278057 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rltxh"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.283361 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n49vg"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.290261 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.290453 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st"] Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.290762 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.790746621 +0000 UTC m=+146.304137924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.310238 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" event={"ID":"3fd1d033-c7b0-44e5-88b8-55d7540079b8","Type":"ContainerStarted","Data":"bfb618b2156c6d7f82e77f048c44303bb63f1ce48a359e092d4a71df69b4ba36"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.310493 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zhqrw"] Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.321026 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s4w7m"] Nov 23 06:52:10 crc kubenswrapper[4906]: W1123 06:52:10.327460 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8b1a645_3eb3_461a_880a_28ab9003a2f0.slice/crio-ffdff691a841ff9d8db794da4b196898d4531c57cf6ae842d916fac26fd8e57c WatchSource:0}: Error finding container ffdff691a841ff9d8db794da4b196898d4531c57cf6ae842d916fac26fd8e57c: Status 404 returned error can't find the container with id ffdff691a841ff9d8db794da4b196898d4531c57cf6ae842d916fac26fd8e57c Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.335099 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" event={"ID":"380e25b3-93bc-4e4c-a82d-2adbb242ae15","Type":"ContainerStarted","Data":"1e6d0bd5c24c9e06ea0395235dbb951b242cd709122820d55b954654a6a74be0"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.342701 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" event={"ID":"0b5efaa1-d4d3-4971-9afb-1490732014d6","Type":"ContainerStarted","Data":"53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.342738 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" event={"ID":"0b5efaa1-d4d3-4971-9afb-1490732014d6","Type":"ContainerStarted","Data":"1be19d1b1c054cf6ad8321ccb6efa014c39d4d959a24ec1496f775dd7ca6affb"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.343150 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.347355 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" event={"ID":"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6","Type":"ContainerStarted","Data":"f96f30388e8ed02eaf2532b5b547b10107f014145351b8d2c57dbd78c650370b"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.349805 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" event={"ID":"2d46f022-7d61-4f91-9834-b8e3670ca17f","Type":"ContainerStarted","Data":"5700bb103ab8b18cd68a7eb84376c6e473d5e06ad03a3d376503303226e5e4a7"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.349826 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" event={"ID":"2d46f022-7d61-4f91-9834-b8e3670ca17f","Type":"ContainerStarted","Data":"f9cbf20ea95eaad5a97a43e42bfca60babf1f9b6799105245e58808ace2b56b0"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.352998 4906 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-zbz6t container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.353053 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" podUID="0b5efaa1-d4d3-4971-9afb-1490732014d6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.360596 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" event={"ID":"b944399b-370e-4acd-9be9-fda8415b0d44","Type":"ContainerStarted","Data":"8db55d8f27de19bcbdcc99e8fbea174157e28664d46a9a19f167cb3f857df030"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.366641 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" podStartSLOduration=125.366621874 podStartE2EDuration="2m5.366621874s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:10.365238487 +0000 UTC m=+145.878629790" watchObservedRunningTime="2025-11-23 06:52:10.366621874 +0000 UTC m=+145.880013177" Nov 23 06:52:10 crc kubenswrapper[4906]: W1123 06:52:10.366864 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd92b911b_cee3_4ccd_8a2c_76dbe0d54fb4.slice/crio-bfa7980b8d07065d140d5a33ebf8eda926c76604de2635c5b78f20de13d833f5 WatchSource:0}: Error finding container bfa7980b8d07065d140d5a33ebf8eda926c76604de2635c5b78f20de13d833f5: Status 404 returned error can't find the container with id bfa7980b8d07065d140d5a33ebf8eda926c76604de2635c5b78f20de13d833f5 Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.379741 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wp8fn" event={"ID":"b7452b46-2b11-4a4f-893e-22bd995151bc","Type":"ContainerStarted","Data":"0d3f3237ef143037775a6ebcbb81cd5e7f7f33915534c43237474f8c2c0a7e80"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.387792 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g7tgv" podStartSLOduration=126.387774758 podStartE2EDuration="2m6.387774758s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:10.385817636 +0000 UTC m=+145.899208939" watchObservedRunningTime="2025-11-23 06:52:10.387774758 +0000 UTC m=+145.901166061" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.392703 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.392827 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.892801162 +0000 UTC m=+146.406192465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.392907 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.393382 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:10.893374707 +0000 UTC m=+146.406766010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.420357 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" event={"ID":"fea909e5-6cef-4949-ad98-a37df3ebe87e","Type":"ContainerStarted","Data":"d1ee180c2e51d8c7df43a4a909db42d325f0896a8e89098606937a2aabeb43f8"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.420443 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.427286 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" event={"ID":"29b40410-6b1e-44ce-98a1-f5148a2ad676","Type":"ContainerStarted","Data":"fe009c8febc6e524721d6d62e368501b6265eb55c9774959ca950c5de8707851"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.427329 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" event={"ID":"29b40410-6b1e-44ce-98a1-f5148a2ad676","Type":"ContainerStarted","Data":"7164e9f92b046ef28d06d91fde399e7cc908a3be647d90b974ea4e7696dcfda1"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.430364 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" event={"ID":"c9efeaef-505f-4243-a2a2-e456fcc85218","Type":"ContainerStarted","Data":"3f2f0ae3d56d45bc6a89016d873e9279e1e41c52539ff3949741bb8cbd2014ce"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.437605 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" event={"ID":"b8016f37-c4f6-4619-8b05-8b32b2f6e443","Type":"ContainerStarted","Data":"af60206e9f0f39e1a1cd256621a9633441116e2a1d370b260d69f057e5826737"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.437656 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" event={"ID":"b8016f37-c4f6-4619-8b05-8b32b2f6e443","Type":"ContainerStarted","Data":"d28f17d8505cd616d832691c3f8fdf7827c6728f62cd9b05b3faf06222689073"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.445992 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" podStartSLOduration=126.44596596 podStartE2EDuration="2m6.44596596s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:10.439413785 +0000 UTC m=+145.952805088" watchObservedRunningTime="2025-11-23 06:52:10.44596596 +0000 UTC m=+145.959357263" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.465293 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" event={"ID":"8e98fdbb-8ec9-4bfc-9875-2cab77551f49","Type":"ContainerStarted","Data":"36efe9b6ef1fa1b6a7697e1e049e94e570390ae7e3f18e88caf29384d2836ec0"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.465380 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" event={"ID":"8e98fdbb-8ec9-4bfc-9875-2cab77551f49","Type":"ContainerStarted","Data":"e2d75d6802d3313ee2d894175a15f7949a7ffef6ed6fbf24140723043354c98d"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.475474 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:10 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:10 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:10 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.475602 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.512820 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c6gcz" podStartSLOduration=126.512800831 podStartE2EDuration="2m6.512800831s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:10.51086371 +0000 UTC m=+146.024255013" watchObservedRunningTime="2025-11-23 06:52:10.512800831 +0000 UTC m=+146.026192134" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.520085 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" event={"ID":"c6b6fbeb-f958-4638-a1ee-bba42048f6ff","Type":"ContainerStarted","Data":"1e91167430253b51cc76005d9aa6f7debe7bbe22ca0a341880a756884bb51351"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.528971 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bdrqj" podStartSLOduration=126.528945822 podStartE2EDuration="2m6.528945822s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:10.481725163 +0000 UTC m=+145.995116466" watchObservedRunningTime="2025-11-23 06:52:10.528945822 +0000 UTC m=+146.042337125" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.534558 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.536176 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.036149004 +0000 UTC m=+146.549540307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.540672 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" event={"ID":"8f49e922-260c-4308-b2f9-e16268dbaf85","Type":"ContainerStarted","Data":"015aa1eae4eabba1befedbdd1bbbae1fc0a25d061079e11905b4724a0271fef8"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.540727 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" event={"ID":"8f49e922-260c-4308-b2f9-e16268dbaf85","Type":"ContainerStarted","Data":"56ffcf792074425cd2810839aa56db1ac3cf20e1e4f3984f8a6060bf2c616535"} Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.573759 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rft95" podStartSLOduration=125.573736446 podStartE2EDuration="2m5.573736446s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:10.566320119 +0000 UTC m=+146.079711422" watchObservedRunningTime="2025-11-23 06:52:10.573736446 +0000 UTC m=+146.087127749" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.639778 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.640400 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.140381903 +0000 UTC m=+146.653773206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.744606 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.745989 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.245966748 +0000 UTC m=+146.759358051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.848300 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.848643 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.348625235 +0000 UTC m=+146.862016538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.903281 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-bfhxf" Nov 23 06:52:10 crc kubenswrapper[4906]: I1123 06:52:10.949274 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:10 crc kubenswrapper[4906]: E1123 06:52:10.949741 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.44971936 +0000 UTC m=+146.963110663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.053412 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.053883 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.553870088 +0000 UTC m=+147.067261391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.161437 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.162035 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.66199133 +0000 UTC m=+147.175382623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.162284 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.163285 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.663254224 +0000 UTC m=+147.176645527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.265820 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.266481 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.766466916 +0000 UTC m=+147.279858219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.371972 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.373548 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.87353069 +0000 UTC m=+147.386921993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.453843 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:11 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:11 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:11 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.454310 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.473924 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.474380 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:11.974356619 +0000 UTC m=+147.487747922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.572907 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" event={"ID":"380e25b3-93bc-4e4c-a82d-2adbb242ae15","Type":"ContainerStarted","Data":"8bf008309d337ebbeab504c18ee3a8d0ffc055b3187ddbcafdd1e032ffbc08fa"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.576327 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" event={"ID":"8a5e4a09-9b16-435b-bc05-00d6d174ede3","Type":"ContainerStarted","Data":"91656ee3229263c7b65d7190070f6c6345f259cf99210c49dfcbd6ba8824d66c"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.579089 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.579548 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.079509152 +0000 UTC m=+147.592900465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.589122 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" event={"ID":"a3ef32c5-fbdd-4689-bbd2-1daae2872cb6","Type":"ContainerStarted","Data":"31c89a2b90263ad6370d0b69428a95747becb544da6f19b0aed037c011b1ac1c"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.590551 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.600769 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-jdr4r" podStartSLOduration=127.600746988 podStartE2EDuration="2m7.600746988s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.599204677 +0000 UTC m=+147.112595980" watchObservedRunningTime="2025-11-23 06:52:11.600746988 +0000 UTC m=+147.114138291" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.602955 4906 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-h4mkr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.603044 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" podUID="a3ef32c5-fbdd-4689-bbd2-1daae2872cb6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.611002 4906 generic.go:334] "Generic (PLEG): container finished" podID="c9efeaef-505f-4243-a2a2-e456fcc85218" containerID="eeeb847448db05e1916b2013cb91e38e8ade3b9a903277a66161fcc7350412da" exitCode=0 Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.611395 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wp8fn" event={"ID":"b7452b46-2b11-4a4f-893e-22bd995151bc","Type":"ContainerStarted","Data":"6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.611433 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" event={"ID":"35486e5f-4cac-4539-8395-8855abb73cf0","Type":"ContainerStarted","Data":"4061661468af893ea6dca6cc9b8838e598188a7e4decc8f80bf1b90c7ff96f88"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.611446 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" event={"ID":"c9efeaef-505f-4243-a2a2-e456fcc85218","Type":"ContainerDied","Data":"eeeb847448db05e1916b2013cb91e38e8ade3b9a903277a66161fcc7350412da"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.622364 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" event={"ID":"b944399b-370e-4acd-9be9-fda8415b0d44","Type":"ContainerStarted","Data":"4749e759253461608e5c4788806072d9010ed2145ea14f22956f5be96e2397f3"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.623350 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.639878 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" podStartSLOduration=126.639851121 podStartE2EDuration="2m6.639851121s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.637251342 +0000 UTC m=+147.150642635" watchObservedRunningTime="2025-11-23 06:52:11.639851121 +0000 UTC m=+147.153242424" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.646437 4906 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8hwdl container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.646509 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" podUID="b944399b-370e-4acd-9be9-fda8415b0d44" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.651669 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" event={"ID":"0a2fd331-4b06-4d1b-94f1-673ea64e9d92","Type":"ContainerStarted","Data":"47a5d2102d69b6505f6affa54c8e34fa7ef2ec809754dab8b357c0e264f0caa2"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.658186 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" event={"ID":"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4","Type":"ContainerStarted","Data":"e5b2a5b84a61e4bfd832cb93b9ffcba1179e59a07dcc491ee2c0d19826e3d971"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.658229 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" event={"ID":"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4","Type":"ContainerStarted","Data":"bfa7980b8d07065d140d5a33ebf8eda926c76604de2635c5b78f20de13d833f5"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.675823 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" podStartSLOduration=126.67579992 podStartE2EDuration="2m6.67579992s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.671602198 +0000 UTC m=+147.184993501" watchObservedRunningTime="2025-11-23 06:52:11.67579992 +0000 UTC m=+147.189191213" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.676377 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" event={"ID":"b8b1a645-3eb3-461a-880a-28ab9003a2f0","Type":"ContainerStarted","Data":"ffdff691a841ff9d8db794da4b196898d4531c57cf6ae842d916fac26fd8e57c"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.684483 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.684912 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.184879011 +0000 UTC m=+147.698270314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.685111 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.686268 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.186258868 +0000 UTC m=+147.699650171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.704798 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" event={"ID":"c6b6fbeb-f958-4638-a1ee-bba42048f6ff","Type":"ContainerStarted","Data":"f5213bb04900ff6edf1f6788b5bdd06b6aad197df458f62aa7afd1f703267999"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.748073 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5txhl" event={"ID":"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381","Type":"ContainerStarted","Data":"6553da97a3ddd2f3b48751cdbebe1e8941a00be9bb2eb11017129df76166133b"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.749025 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wp8fn" podStartSLOduration=127.749000681 podStartE2EDuration="2m7.749000681s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.748816286 +0000 UTC m=+147.262207589" watchObservedRunningTime="2025-11-23 06:52:11.749000681 +0000 UTC m=+147.262391984" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.775489 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" event={"ID":"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b","Type":"ContainerStarted","Data":"2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.775556 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" event={"ID":"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b","Type":"ContainerStarted","Data":"31c9724b8f16bbd939f7e535354e71d9f58d0ae97b910f72b28a615ee4a46a3c"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.776899 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.777910 4906 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8ktpz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.777977 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.790525 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c62nz" podStartSLOduration=127.790501208 podStartE2EDuration="2m7.790501208s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.776061303 +0000 UTC m=+147.289452616" watchObservedRunningTime="2025-11-23 06:52:11.790501208 +0000 UTC m=+147.303892511" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.800220 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.802146 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.302120247 +0000 UTC m=+147.815511550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.802584 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.805487 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.305460037 +0000 UTC m=+147.818851340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.847444 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" event={"ID":"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d","Type":"ContainerStarted","Data":"7a483f10bfdb3e776ba543e4666b1f2c56ae3d4c38cb4bdb5a64ff1f4198c783"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.847480 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" event={"ID":"2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d","Type":"ContainerStarted","Data":"12bdf97218db8d04de641d4a95d0a5a45de14e33bda5c82c982135eb36ed16b0"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.849467 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" podStartSLOduration=127.849424909 podStartE2EDuration="2m7.849424909s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.837670406 +0000 UTC m=+147.351061709" watchObservedRunningTime="2025-11-23 06:52:11.849424909 +0000 UTC m=+147.362816212" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.849563 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.849657 4906 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tkds2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.849743 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" podUID="2c1cfb8a-1d8f-44b8-9c99-c915dcc4398d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.868366 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" event={"ID":"803fc706-4690-447f-9aa8-bb512d9136f2","Type":"ContainerStarted","Data":"95cd6de3ced2e6c5e0ab19496c4e5692ce224811f8e4f0359117946c96139c1e"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.878626 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" podStartSLOduration=126.878608017 podStartE2EDuration="2m6.878608017s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.877925278 +0000 UTC m=+147.391316591" watchObservedRunningTime="2025-11-23 06:52:11.878608017 +0000 UTC m=+147.391999320" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.889342 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" event={"ID":"be5a3686-7427-4990-b4ec-6944a80bab27","Type":"ContainerStarted","Data":"15bd8954ee60d696838f141fbcd249e4b2a074c3c6c0219e194d4a9d92a4f249"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.889484 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" event={"ID":"be5a3686-7427-4990-b4ec-6944a80bab27","Type":"ContainerStarted","Data":"4bb143a9cd373831af738660ba7ddb5f34a398ce921abd80b994c70a55221f6a"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.898125 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s4w7m" event={"ID":"6255e00b-ec87-429b-bc92-1df3cb176a5e","Type":"ContainerStarted","Data":"200b69023796ee83ac54e21a4e3083260971206aaa575b86d9b102b90a1ccbf9"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.898204 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s4w7m" event={"ID":"6255e00b-ec87-429b-bc92-1df3cb176a5e","Type":"ContainerStarted","Data":"6b4c7bb5c2ba719215b6e0d7046b659726340cbeb720b4eb61d6caf5cbccdd3d"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.900794 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" podStartSLOduration=126.900773518 podStartE2EDuration="2m6.900773518s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.897525701 +0000 UTC m=+147.410916994" watchObservedRunningTime="2025-11-23 06:52:11.900773518 +0000 UTC m=+147.414164821" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.906271 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.406247474 +0000 UTC m=+147.919638777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.904143 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.908850 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:11 crc kubenswrapper[4906]: E1123 06:52:11.910016 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.409998934 +0000 UTC m=+147.923390237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.910999 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" event={"ID":"51d28f4b-a6a1-4698-9081-52eb141bd832","Type":"ContainerStarted","Data":"6d686ebf3857a7a6f8909522a5634442f0f3f003e2e9a71607c8e777d4f81031"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.912260 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" event={"ID":"51d28f4b-a6a1-4698-9081-52eb141bd832","Type":"ContainerStarted","Data":"fd48b886e7be0752a71612b68655c9600d854fd63a30992b45019a8ff12a4bc7"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.928217 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" event={"ID":"3fd1d033-c7b0-44e5-88b8-55d7540079b8","Type":"ContainerStarted","Data":"9a8f40e5e31ac1e55d425e3068f5f163cc94968e3c25359055db1b556625cf24"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.928279 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" event={"ID":"3fd1d033-c7b0-44e5-88b8-55d7540079b8","Type":"ContainerStarted","Data":"c6805f31f07e6fdf74e4e015afae86f299177905b5c42e1bdf39d69f12d27aaf"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.932468 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" event={"ID":"c490ab53-76a1-473e-b007-2493b5dc2f2a","Type":"ContainerStarted","Data":"898d134ba528515da0460232d115413bff8827c0f04ff0a15f05df4c8955f553"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.932502 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" event={"ID":"c490ab53-76a1-473e-b007-2493b5dc2f2a","Type":"ContainerStarted","Data":"cfbdb27a4c4d1314f31ad5d5576f76af492d08a9d2147daf88277500afbd199c"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.954625 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" podStartSLOduration=127.954602413 podStartE2EDuration="2m7.954602413s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.925176408 +0000 UTC m=+147.438567711" watchObservedRunningTime="2025-11-23 06:52:11.954602413 +0000 UTC m=+147.467993706" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.978126 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" event={"ID":"87eb0e66-14a8-4c87-8235-986e5bccec2b","Type":"ContainerStarted","Data":"99ceb01d57deadbdf8612675d6670fbd4140c018dc41e4c9cb2739dbf41eca09"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.978400 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" event={"ID":"87eb0e66-14a8-4c87-8235-986e5bccec2b","Type":"ContainerStarted","Data":"52118795251c094a14f3d8a842c9e3bd85300ed6012dc648160f880a5ee2cf73"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.978581 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" event={"ID":"87eb0e66-14a8-4c87-8235-986e5bccec2b","Type":"ContainerStarted","Data":"d618401c0d6ef18d957eef9bb7e60c3118651344842b1f86ef0cdf79cc04e3e1"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.991457 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-s4w7m" podStartSLOduration=7.991432615 podStartE2EDuration="7.991432615s" podCreationTimestamp="2025-11-23 06:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.957060439 +0000 UTC m=+147.470451742" watchObservedRunningTime="2025-11-23 06:52:11.991432615 +0000 UTC m=+147.504823918" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.994486 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" event={"ID":"b8016f37-c4f6-4619-8b05-8b32b2f6e443","Type":"ContainerStarted","Data":"2f0e7587de95e40b78e62b0b05fb86e8c0262642969dc377572b0358fa9eaad3"} Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.994870 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-vhv4d" podStartSLOduration=127.994863867 podStartE2EDuration="2m7.994863867s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:11.980928225 +0000 UTC m=+147.494319528" watchObservedRunningTime="2025-11-23 06:52:11.994863867 +0000 UTC m=+147.508255170" Nov 23 06:52:11 crc kubenswrapper[4906]: I1123 06:52:11.995430 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.008636 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tgwrc" podStartSLOduration=128.008617144 podStartE2EDuration="2m8.008617144s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:12.007177605 +0000 UTC m=+147.520568908" watchObservedRunningTime="2025-11-23 06:52:12.008617144 +0000 UTC m=+147.522008447" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.010667 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.012340 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.512300631 +0000 UTC m=+148.025691934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.016751 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" event={"ID":"dc8a7161-8582-4a46-a920-a8c635957b0f","Type":"ContainerStarted","Data":"2688b357b619c3a1e0ae952803f7c42b8fbba897debf94296649d482fd5a6e54"} Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.016893 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" event={"ID":"dc8a7161-8582-4a46-a920-a8c635957b0f","Type":"ContainerStarted","Data":"241bb276389fd88af67b0600025c8689ac1dd4f002bebe6c4cd5d19119af533f"} Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.028076 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-65gnv" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.069801 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" podStartSLOduration=127.069767843 podStartE2EDuration="2m7.069767843s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:12.046987386 +0000 UTC m=+147.560378689" watchObservedRunningTime="2025-11-23 06:52:12.069767843 +0000 UTC m=+147.583159146" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.072893 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.075107 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.076079 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.085597 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.085886 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.086164 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.086198 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cfqk7" podStartSLOduration=128.086171951 podStartE2EDuration="2m8.086171951s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:12.07710807 +0000 UTC m=+147.590499373" watchObservedRunningTime="2025-11-23 06:52:12.086171951 +0000 UTC m=+147.599563254" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.116887 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.119512 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.119663 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.124846 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.620251579 +0000 UTC m=+148.133642882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.198724 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rltxh" podStartSLOduration=127.198700491 podStartE2EDuration="2m7.198700491s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:12.194953402 +0000 UTC m=+147.708344705" watchObservedRunningTime="2025-11-23 06:52:12.198700491 +0000 UTC m=+147.712091794" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.228337 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.228542 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.228595 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.228774 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.228995 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.728972718 +0000 UTC m=+148.242364021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.270928 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.339513 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.340467 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.840444331 +0000 UTC m=+148.353835634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.361281 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.407507 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.442278 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.442711 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.942694117 +0000 UTC m=+148.456085420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.442824 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.444277 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:12.944253209 +0000 UTC m=+148.457644512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.461935 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:12 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:12 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:12 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.461999 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.547571 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.548329 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.048301112 +0000 UTC m=+148.561692415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.649843 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.650500 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.150466236 +0000 UTC m=+148.663857539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.751404 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.752271 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.25224722 +0000 UTC m=+148.765638523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.853557 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.854027 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.354008773 +0000 UTC m=+148.867400076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.957298 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:12 crc kubenswrapper[4906]: E1123 06:52:12.958826 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.458777317 +0000 UTC m=+148.972168620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:12 crc kubenswrapper[4906]: I1123 06:52:12.971462 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.059619 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.061304 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.56128825 +0000 UTC m=+149.074679553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.066417 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h48nj" event={"ID":"be5a3686-7427-4990-b4ec-6944a80bab27","Type":"ContainerStarted","Data":"7647a77d4fe8571480f024429ff1a4739176c3d76941d5f78e81bc6152e2769f"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.073673 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" event={"ID":"803fc706-4690-447f-9aa8-bb512d9136f2","Type":"ContainerStarted","Data":"e0652494a9087622e43571ad822143c2630cd01c43165e330a02736573aae8bb"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.075104 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.085063 4906 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8hr2w container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.31:6443/healthz\": dial tcp 10.217.0.31:6443: connect: connection refused" start-of-body= Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.085156 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" podUID="803fc706-4690-447f-9aa8-bb512d9136f2" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.31:6443/healthz\": dial tcp 10.217.0.31:6443: connect: connection refused" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.095273 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5txhl" event={"ID":"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381","Type":"ContainerStarted","Data":"e170b56e603910bd73070cd8e4d3d0796114d256f979559637125c5c7971fb56"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.095344 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5txhl" event={"ID":"4ad26aa7-4af7-4f02-87c6-6a6ce70bd381","Type":"ContainerStarted","Data":"3b79493ac358dd1635e7274c11f736becbc9c8f2b2bfcf650d65f51884089103"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.096389 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.103743 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" event={"ID":"c9efeaef-505f-4243-a2a2-e456fcc85218","Type":"ContainerStarted","Data":"02998351c1f4a3d2c5c7886f80c38b36f21d13cc1692141b91ce39754e0ba0a8"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.134884 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" event={"ID":"35486e5f-4cac-4539-8395-8855abb73cf0","Type":"ContainerStarted","Data":"44495fc1164d5b4c240ffac5d7f33cce15a82d5596c2f254e68095bd0391edc3"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.134936 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" event={"ID":"35486e5f-4cac-4539-8395-8855abb73cf0","Type":"ContainerStarted","Data":"0baf4653354a9883f12e8ed03bf044efd663b34b89c4c919961ec677833a4f49"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.137665 4906 generic.go:334] "Generic (PLEG): container finished" podID="0a2fd331-4b06-4d1b-94f1-673ea64e9d92" containerID="d5adae9de8bbe814c3c714f68374696584468505e23de99ddcfa51606772d0d2" exitCode=0 Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.137768 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" event={"ID":"0a2fd331-4b06-4d1b-94f1-673ea64e9d92","Type":"ContainerDied","Data":"d5adae9de8bbe814c3c714f68374696584468505e23de99ddcfa51606772d0d2"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.147507 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" event={"ID":"b8b1a645-3eb3-461a-880a-28ab9003a2f0","Type":"ContainerStarted","Data":"dd1b5e5304d95e48f62640cd34c12fa90da63869cb270370a93f9c83b4c3c182"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.147581 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" event={"ID":"b8b1a645-3eb3-461a-880a-28ab9003a2f0","Type":"ContainerStarted","Data":"762da95592159676b2b9893981b339feaed1b6cf19feef9209d8d0e969cebddf"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.149577 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-5txhl" podStartSLOduration=9.149555614 podStartE2EDuration="9.149555614s" podCreationTimestamp="2025-11-23 06:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:13.148704301 +0000 UTC m=+148.662095604" watchObservedRunningTime="2025-11-23 06:52:13.149555614 +0000 UTC m=+148.662946917" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.150846 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" podStartSLOduration=129.150838427 podStartE2EDuration="2m9.150838427s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:13.11904213 +0000 UTC m=+148.632433433" watchObservedRunningTime="2025-11-23 06:52:13.150838427 +0000 UTC m=+148.664229730" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.163112 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.165012 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.664964885 +0000 UTC m=+149.178356198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.184941 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" event={"ID":"8a5e4a09-9b16-435b-bc05-00d6d174ede3","Type":"ContainerStarted","Data":"11c8582c632b4e1612805331247b8be1c29402338358ad179344d293b1facf2a"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.186556 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" event={"ID":"51d28f4b-a6a1-4698-9081-52eb141bd832","Type":"ContainerStarted","Data":"b3b6dbf106393d5981fae32ad15da3a8ef41c31f23fca627d3bffb86dd9e0de5"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.209875 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"26f0a9bc-3f7f-4005-8775-6f85f2db24e5","Type":"ContainerStarted","Data":"8667ddb2aa219f536e5d9eda554eada9ae8c5813bdef2926ff7f7a2d5805469b"} Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.211225 4906 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8ktpz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.211304 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.216740 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-h4mkr" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.224125 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8hwdl" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.229656 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rwvsq" podStartSLOduration=129.229626589 podStartE2EDuration="2m9.229626589s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:13.228885278 +0000 UTC m=+148.742276581" watchObservedRunningTime="2025-11-23 06:52:13.229626589 +0000 UTC m=+148.743017892" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.238316 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tkds2" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.267603 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.268046 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.275819 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.275870 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.275939 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.283066 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.783048583 +0000 UTC m=+149.296439886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.285882 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-n49vg" podStartSLOduration=128.285856468 podStartE2EDuration="2m8.285856468s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:13.282515778 +0000 UTC m=+148.795907081" watchObservedRunningTime="2025-11-23 06:52:13.285856468 +0000 UTC m=+148.799247771" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.289476 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.290098 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.291357 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.298731 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.381626 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.388060 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.888037372 +0000 UTC m=+149.401428665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.392947 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.408420 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.420360 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.448351 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:13 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:13 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:13 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.448442 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.467045 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vppvc" podStartSLOduration=129.467021788 podStartE2EDuration="2m9.467021788s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:13.464937482 +0000 UTC m=+148.978328775" watchObservedRunningTime="2025-11-23 06:52:13.467021788 +0000 UTC m=+148.980413091" Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.483031 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.483360 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:13.983347033 +0000 UTC m=+149.496738336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.591593 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.591841 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.091813095 +0000 UTC m=+149.605204398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.592251 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.592561 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.092548785 +0000 UTC m=+149.605940088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.692616 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.692869 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.192852469 +0000 UTC m=+149.706243762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.793940 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.794692 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.294663803 +0000 UTC m=+149.808055106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:13 crc kubenswrapper[4906]: I1123 06:52:13.895127 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:13 crc kubenswrapper[4906]: E1123 06:52:13.895478 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.395460371 +0000 UTC m=+149.908851674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.000934 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.001531 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.501510349 +0000 UTC m=+150.014901652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.101971 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.102501 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.60244585 +0000 UTC m=+150.115837153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.104576 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.105160 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.605145982 +0000 UTC m=+150.118537285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.207954 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.208227 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.70821033 +0000 UTC m=+150.221601633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.252010 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ln24p"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.255272 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.265084 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.267574 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"650f8191f738df2231f9f4e1f49df8c8ff3b1b055470f2c7b752d0053c1dd9df"} Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.271983 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"26f0a9bc-3f7f-4005-8775-6f85f2db24e5","Type":"ContainerStarted","Data":"e17b42bb4309e4b9587fc31ac57bcb282a09fc1a03de76835a59d817867e455e"} Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.277116 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" event={"ID":"c9efeaef-505f-4243-a2a2-e456fcc85218","Type":"ContainerStarted","Data":"4257944d5b5499034079926d47560397d7ebbeef662b2d0bf30941e45a8eb131"} Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.287225 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"18bfe1aeab941acd9bd458adca8a8ce440e28ead431322c153d914337d4b8a0b"} Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.296451 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" event={"ID":"0a2fd331-4b06-4d1b-94f1-673ea64e9d92","Type":"ContainerStarted","Data":"0438ea1486a3bcc4d5b09fd2bc477db175c32965fa04db42aa6bf0593503b8e6"} Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.298178 4906 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8ktpz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.298222 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.299369 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ln24p"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.311466 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-utilities\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.311672 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-catalog-content\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.311824 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4gkq\" (UniqueName: \"kubernetes.io/projected/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-kube-api-access-f4gkq\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.311964 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.316897 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.322037 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.822018054 +0000 UTC m=+150.335409357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.378689 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" podStartSLOduration=129.378655214 podStartE2EDuration="2m9.378655214s" podCreationTimestamp="2025-11-23 06:50:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:14.377425641 +0000 UTC m=+149.890816944" watchObservedRunningTime="2025-11-23 06:52:14.378655214 +0000 UTC m=+149.892046507" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.413586 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.413794 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.91376466 +0000 UTC m=+150.427155963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.413870 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-catalog-content\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.413923 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4gkq\" (UniqueName: \"kubernetes.io/projected/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-kube-api-access-f4gkq\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.413983 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.414032 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-utilities\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.414355 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-catalog-content\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.415226 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-utilities\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.416986 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:14.916976566 +0000 UTC m=+150.430367869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.460133 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4gkq\" (UniqueName: \"kubernetes.io/projected/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-kube-api-access-f4gkq\") pod \"certified-operators-ln24p\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.460235 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:14 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:14 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:14 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.460288 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.462232 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pltlf"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.463279 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.469074 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.496342 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pltlf"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.501143 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" podStartSLOduration=130.50112046 podStartE2EDuration="2m10.50112046s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:14.489625043 +0000 UTC m=+150.003016346" watchObservedRunningTime="2025-11-23 06:52:14.50112046 +0000 UTC m=+150.014511763" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.521019 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.521537 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjtlf\" (UniqueName: \"kubernetes.io/projected/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-kube-api-access-wjtlf\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.521609 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-utilities\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.521672 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.021642736 +0000 UTC m=+150.535034039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.521821 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-catalog-content\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.525308 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.525296594 podStartE2EDuration="3.525296594s" podCreationTimestamp="2025-11-23 06:52:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:14.51727132 +0000 UTC m=+150.030662643" watchObservedRunningTime="2025-11-23 06:52:14.525296594 +0000 UTC m=+150.038687897" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.623182 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-utilities\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.623257 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-catalog-content\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.623299 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.623342 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjtlf\" (UniqueName: \"kubernetes.io/projected/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-kube-api-access-wjtlf\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.623875 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.123858351 +0000 UTC m=+150.637249654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.623887 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-utilities\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.623937 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-catalog-content\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.641897 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zsl6g"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.645399 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.668531 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zsl6g"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.684921 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjtlf\" (UniqueName: \"kubernetes.io/projected/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-kube-api-access-wjtlf\") pod \"community-operators-pltlf\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.689219 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.725060 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.725209 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.225186224 +0000 UTC m=+150.738577527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.727334 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.22731014 +0000 UTC m=+150.740701443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.727269 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.727663 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-utilities\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.727834 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-catalog-content\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.728916 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8pc5\" (UniqueName: \"kubernetes.io/projected/c80737c3-e70d-40eb-a8ed-d3f4dba27233-kube-api-access-r8pc5\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.817481 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.832833 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.833074 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8pc5\" (UniqueName: \"kubernetes.io/projected/c80737c3-e70d-40eb-a8ed-d3f4dba27233-kube-api-access-r8pc5\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.833154 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-utilities\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.833169 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-catalog-content\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.834039 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-catalog-content\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.834119 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.334102108 +0000 UTC m=+150.847493411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.834505 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-utilities\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.842061 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9xlcz"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.843388 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.851596 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9xlcz"] Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.888936 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8pc5\" (UniqueName: \"kubernetes.io/projected/c80737c3-e70d-40eb-a8ed-d3f4dba27233-kube-api-access-r8pc5\") pod \"certified-operators-zsl6g\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.938987 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.939045 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-utilities\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.939093 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqfwv\" (UniqueName: \"kubernetes.io/projected/aef8b5df-b818-47b6-9d3e-8266b55d3be7-kube-api-access-kqfwv\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.939159 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-catalog-content\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:14 crc kubenswrapper[4906]: E1123 06:52:14.939477 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.439462147 +0000 UTC m=+150.952853450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:14 crc kubenswrapper[4906]: I1123 06:52:14.993140 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.031008 4906 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.038559 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ln24p"] Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.039879 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.040025 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-utilities\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.040063 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqfwv\" (UniqueName: \"kubernetes.io/projected/aef8b5df-b818-47b6-9d3e-8266b55d3be7-kube-api-access-kqfwv\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.040100 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-catalog-content\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.040531 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-catalog-content\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.040593 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.540578752 +0000 UTC m=+151.053970055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.040816 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-utilities\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.057118 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqfwv\" (UniqueName: \"kubernetes.io/projected/aef8b5df-b818-47b6-9d3e-8266b55d3be7-kube-api-access-kqfwv\") pod \"community-operators-9xlcz\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:15 crc kubenswrapper[4906]: W1123 06:52:15.088916 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6eaabb7_44d7_4502_958d_e2a2ed1b06ea.slice/crio-f169430360d156e7dcf65537d7813aa7299fa0bbd7015c695c4106f6cc87a80c WatchSource:0}: Error finding container f169430360d156e7dcf65537d7813aa7299fa0bbd7015c695c4106f6cc87a80c: Status 404 returned error can't find the container with id f169430360d156e7dcf65537d7813aa7299fa0bbd7015c695c4106f6cc87a80c Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.146020 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.147250 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.647232276 +0000 UTC m=+151.160623579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.198972 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pltlf"] Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.206410 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.249531 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.249896 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.749868733 +0000 UTC m=+151.263260036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.309986 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pltlf" event={"ID":"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04","Type":"ContainerStarted","Data":"64b1ca4355d513f73d2bac44905aae2233586400eec33ce15dd3a8e62f57eba4"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.311276 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerStarted","Data":"8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.311297 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerStarted","Data":"f169430360d156e7dcf65537d7813aa7299fa0bbd7015c695c4106f6cc87a80c"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.330011 4906 generic.go:334] "Generic (PLEG): container finished" podID="d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" containerID="e5b2a5b84a61e4bfd832cb93b9ffcba1179e59a07dcc491ee2c0d19826e3d971" exitCode=0 Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.330131 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" event={"ID":"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4","Type":"ContainerDied","Data":"e5b2a5b84a61e4bfd832cb93b9ffcba1179e59a07dcc491ee2c0d19826e3d971"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.351879 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" event={"ID":"8a5e4a09-9b16-435b-bc05-00d6d174ede3","Type":"ContainerStarted","Data":"6080b8e7f89466946965c9a1b7d083910f566f1a4b9d337b3cee5c6aa6a8775b"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.352239 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" event={"ID":"8a5e4a09-9b16-435b-bc05-00d6d174ede3","Type":"ContainerStarted","Data":"6cac8faacb4f06dd8c20c7175613cb01e467bad74118259e21aa977a69377323"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.353807 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7830aefb96ac93772b53e7a857d4f4f99d0cedc0b38ede576b6fab4e1c2bf997"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.354032 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.355867 4906 generic.go:334] "Generic (PLEG): container finished" podID="26f0a9bc-3f7f-4005-8775-6f85f2db24e5" containerID="e17b42bb4309e4b9587fc31ac57bcb282a09fc1a03de76835a59d817867e455e" exitCode=0 Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.361702 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.362304 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.86227852 +0000 UTC m=+151.375669813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.365828 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"26f0a9bc-3f7f-4005-8775-6f85f2db24e5","Type":"ContainerDied","Data":"e17b42bb4309e4b9587fc31ac57bcb282a09fc1a03de76835a59d817867e455e"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.365877 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f8fdefa61271826d73805f2a5011cf62f1a827affe0dbfdf42547d5131d4af0e"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.365890 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6f3ffc2a51cf02ccc1d5bc40b15e84fb9d4c6ae386f33ba5bcfcccf93ffc65a0"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.365902 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2ee016fa701094d1f45113b86e9be942270f9af2889b18e586f2be919d5d2147"} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.454785 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:15 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:15 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:15 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.454842 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.462932 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.463173 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.963122749 +0000 UTC m=+151.476514042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.463367 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.468771 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:15.968752479 +0000 UTC m=+151.482143972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.543664 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zsl6g"] Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.585439 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.585590 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:52:16.085562783 +0000 UTC m=+151.598954086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.585804 4906 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-23T06:52:15.031037208Z","Handler":null,"Name":""} Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.593742 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:15 crc kubenswrapper[4906]: E1123 06:52:15.594044 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:52:16.094030358 +0000 UTC m=+151.607421661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ll4c9" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.597357 4906 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.597406 4906 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.644088 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9xlcz"] Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.696301 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.724109 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.724989 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.732200 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.732408 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.735630 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.737991 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.798140 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.798290 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/783a71d5-441c-49be-aad7-64d3038cac6b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.798363 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/783a71d5-441c-49be-aad7-64d3038cac6b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.807598 4906 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.808091 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.844201 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ll4c9\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.900067 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/783a71d5-441c-49be-aad7-64d3038cac6b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.900195 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/783a71d5-441c-49be-aad7-64d3038cac6b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.900273 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/783a71d5-441c-49be-aad7-64d3038cac6b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:15 crc kubenswrapper[4906]: I1123 06:52:15.921963 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/783a71d5-441c-49be-aad7-64d3038cac6b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.028058 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.067723 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.237955 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zxrrv"] Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.239424 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.242921 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.257755 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxrrv"] Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.264578 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ll4c9"] Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.307605 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-utilities\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.307645 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dmf7\" (UniqueName: \"kubernetes.io/projected/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-kube-api-access-9dmf7\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.307730 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-catalog-content\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.317387 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 06:52:16 crc kubenswrapper[4906]: W1123 06:52:16.334567 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod783a71d5_441c_49be_aad7_64d3038cac6b.slice/crio-8aae5d1658eedfcb105f9b586c225acec92bb6b8f95e46d9c75b675b966f85b2 WatchSource:0}: Error finding container 8aae5d1658eedfcb105f9b586c225acec92bb6b8f95e46d9c75b675b966f85b2: Status 404 returned error can't find the container with id 8aae5d1658eedfcb105f9b586c225acec92bb6b8f95e46d9c75b675b966f85b2 Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.369966 4906 generic.go:334] "Generic (PLEG): container finished" podID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerID="58822ed64ecfcd73e5382d685cf6edbaa2b6c1dcc2d7f051bed3c69b95584949" exitCode=0 Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.370076 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xlcz" event={"ID":"aef8b5df-b818-47b6-9d3e-8266b55d3be7","Type":"ContainerDied","Data":"58822ed64ecfcd73e5382d685cf6edbaa2b6c1dcc2d7f051bed3c69b95584949"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.370116 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xlcz" event={"ID":"aef8b5df-b818-47b6-9d3e-8266b55d3be7","Type":"ContainerStarted","Data":"bde44e6abf72230599539ea020e593d4940410a15c3cc018a9e281c3bba502c4"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.373862 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.375811 4906 generic.go:334] "Generic (PLEG): container finished" podID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerID="0d2c6987dcaa78211881fba491844ee05032a34164a9a6acaf6548507c7463f9" exitCode=0 Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.375929 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsl6g" event={"ID":"c80737c3-e70d-40eb-a8ed-d3f4dba27233","Type":"ContainerDied","Data":"0d2c6987dcaa78211881fba491844ee05032a34164a9a6acaf6548507c7463f9"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.375966 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsl6g" event={"ID":"c80737c3-e70d-40eb-a8ed-d3f4dba27233","Type":"ContainerStarted","Data":"e32262f90ed4869256840ba2ef6ebac59daec419d35d7ac06d42c1162c900574"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.390870 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"783a71d5-441c-49be-aad7-64d3038cac6b","Type":"ContainerStarted","Data":"8aae5d1658eedfcb105f9b586c225acec92bb6b8f95e46d9c75b675b966f85b2"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.394909 4906 generic.go:334] "Generic (PLEG): container finished" podID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerID="f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675" exitCode=0 Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.395458 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pltlf" event={"ID":"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04","Type":"ContainerDied","Data":"f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.402109 4906 generic.go:334] "Generic (PLEG): container finished" podID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerID="8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572" exitCode=0 Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.402216 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerDied","Data":"8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.409181 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-utilities\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.409213 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dmf7\" (UniqueName: \"kubernetes.io/projected/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-kube-api-access-9dmf7\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.409247 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-catalog-content\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.410120 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" event={"ID":"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a","Type":"ContainerStarted","Data":"854f5dbb0197b59c1c73e548bb61162172dbf14ca782defe2e5691e71b1b742b"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.410166 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-catalog-content\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.410377 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-utilities\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.415092 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" event={"ID":"8a5e4a09-9b16-435b-bc05-00d6d174ede3","Type":"ContainerStarted","Data":"f8a7c383eb884c100997ba70c3f13dd4debb9776808bcdb2378424bf3fad146d"} Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.438792 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dmf7\" (UniqueName: \"kubernetes.io/projected/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-kube-api-access-9dmf7\") pod \"redhat-marketplace-zxrrv\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.444391 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.460902 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" podStartSLOduration=12.46087702 podStartE2EDuration="12.46087702s" podCreationTimestamp="2025-11-23 06:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:16.447301619 +0000 UTC m=+151.960692922" watchObservedRunningTime="2025-11-23 06:52:16.46087702 +0000 UTC m=+151.974268323" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.465533 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:16 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:16 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:16 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.465610 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.559490 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.654474 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-49hxc"] Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.656907 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.658590 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-49hxc"] Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.660436 4906 patch_prober.go:28] interesting pod/downloads-7954f5f757-4qmrn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.660508 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4qmrn" podUID="15d81e07-ddd3-436e-9848-14f986e9381f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.660932 4906 patch_prober.go:28] interesting pod/downloads-7954f5f757-4qmrn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.660955 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-4qmrn" podUID="15d81e07-ddd3-436e-9848-14f986e9381f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.735607 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-utilities\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.736346 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-catalog-content\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.736411 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt64t\" (UniqueName: \"kubernetes.io/projected/47ebbb68-ad04-4889-b010-de1719967aa7-kube-api-access-mt64t\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.792936 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.837695 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-utilities\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.837777 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-catalog-content\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.838402 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt64t\" (UniqueName: \"kubernetes.io/projected/47ebbb68-ad04-4889-b010-de1719967aa7-kube-api-access-mt64t\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.839245 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-utilities\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.839866 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-catalog-content\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.845293 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.871003 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt64t\" (UniqueName: \"kubernetes.io/projected/47ebbb68-ad04-4889-b010-de1719967aa7-kube-api-access-mt64t\") pod \"redhat-marketplace-49hxc\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.940544 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-config-volume\") pod \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.940630 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22d56\" (UniqueName: \"kubernetes.io/projected/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-kube-api-access-22d56\") pod \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.940666 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kubelet-dir\") pod \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.940708 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-secret-volume\") pod \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\" (UID: \"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4\") " Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.940751 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kube-api-access\") pod \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\" (UID: \"26f0a9bc-3f7f-4005-8775-6f85f2db24e5\") " Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.940977 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "26f0a9bc-3f7f-4005-8775-6f85f2db24e5" (UID: "26f0a9bc-3f7f-4005-8775-6f85f2db24e5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.942369 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-config-volume" (OuterVolumeSpecName: "config-volume") pod "d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" (UID: "d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.945111 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "26f0a9bc-3f7f-4005-8775-6f85f2db24e5" (UID: "26f0a9bc-3f7f-4005-8775-6f85f2db24e5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.946402 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-kube-api-access-22d56" (OuterVolumeSpecName: "kube-api-access-22d56") pod "d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" (UID: "d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4"). InnerVolumeSpecName "kube-api-access-22d56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.946590 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" (UID: "d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:52:16 crc kubenswrapper[4906]: I1123 06:52:16.990495 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.042190 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.042229 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22d56\" (UniqueName: \"kubernetes.io/projected/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-kube-api-access-22d56\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.042241 4906 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.042248 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.042257 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/26f0a9bc-3f7f-4005-8775-6f85f2db24e5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.285212 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxrrv"] Nov 23 06:52:17 crc kubenswrapper[4906]: W1123 06:52:17.320037 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dc395cc_508a_42a1_ab5a_06cc4d3f0167.slice/crio-31ba81d7e03dd37cc4a8550024b4b8d5339c1181e602c736f6cea8f4fef4250d WatchSource:0}: Error finding container 31ba81d7e03dd37cc4a8550024b4b8d5339c1181e602c736f6cea8f4fef4250d: Status 404 returned error can't find the container with id 31ba81d7e03dd37cc4a8550024b4b8d5339c1181e602c736f6cea8f4fef4250d Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.372537 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.375569 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-49hxc"] Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.436042 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.436290 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.446800 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.449516 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"26f0a9bc-3f7f-4005-8775-6f85f2db24e5","Type":"ContainerDied","Data":"8667ddb2aa219f536e5d9eda554eada9ae8c5813bdef2926ff7f7a2d5805469b"} Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.449573 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8667ddb2aa219f536e5d9eda554eada9ae8c5813bdef2926ff7f7a2d5805469b" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.449672 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.453260 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:17 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:17 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:17 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.453348 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.467969 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxrrv" event={"ID":"0dc395cc-508a-42a1-ab5a-06cc4d3f0167","Type":"ContainerStarted","Data":"31ba81d7e03dd37cc4a8550024b4b8d5339c1181e602c736f6cea8f4fef4250d"} Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.481366 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.481777 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"783a71d5-441c-49be-aad7-64d3038cac6b","Type":"ContainerStarted","Data":"8c9594c130c242c0fd021c766415c6abb1203f7ed157a2a841696c83840b1d46"} Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.499790 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" event={"ID":"d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4","Type":"ContainerDied","Data":"bfa7980b8d07065d140d5a33ebf8eda926c76604de2635c5b78f20de13d833f5"} Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.499833 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfa7980b8d07065d140d5a33ebf8eda926c76604de2635c5b78f20de13d833f5" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.499897 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.530016 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" event={"ID":"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a","Type":"ContainerStarted","Data":"ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309"} Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.530062 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.622763 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" podStartSLOduration=133.622739789 podStartE2EDuration="2m13.622739789s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:17.618841126 +0000 UTC m=+153.132232429" watchObservedRunningTime="2025-11-23 06:52:17.622739789 +0000 UTC m=+153.136131092" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.637184 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r2vnv"] Nov 23 06:52:17 crc kubenswrapper[4906]: E1123 06:52:17.638550 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" containerName="collect-profiles" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.640364 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" containerName="collect-profiles" Nov 23 06:52:17 crc kubenswrapper[4906]: E1123 06:52:17.640471 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f0a9bc-3f7f-4005-8775-6f85f2db24e5" containerName="pruner" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.640550 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f0a9bc-3f7f-4005-8775-6f85f2db24e5" containerName="pruner" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.640795 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f0a9bc-3f7f-4005-8775-6f85f2db24e5" containerName="pruner" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.640883 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" containerName="collect-profiles" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.642477 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.645480 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.661875 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r2vnv"] Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.766741 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-utilities\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.766807 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-catalog-content\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.767007 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z8vb\" (UniqueName: \"kubernetes.io/projected/7147ecc9-6043-4838-9c47-762052b68bf1-kube-api-access-8z8vb\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.868257 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z8vb\" (UniqueName: \"kubernetes.io/projected/7147ecc9-6043-4838-9c47-762052b68bf1-kube-api-access-8z8vb\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.868354 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-utilities\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.868384 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-catalog-content\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.868904 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-catalog-content\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.869269 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-utilities\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:17 crc kubenswrapper[4906]: I1123 06:52:17.904335 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z8vb\" (UniqueName: \"kubernetes.io/projected/7147ecc9-6043-4838-9c47-762052b68bf1-kube-api-access-8z8vb\") pod \"redhat-operators-r2vnv\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.041554 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8qsds"] Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.042759 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.068333 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.068407 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.069145 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8qsds"] Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.078262 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.088169 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.121933 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.122725 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.123921 4906 patch_prober.go:28] interesting pod/console-f9d7485db-wp8fn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.123961 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wp8fn" podUID="b7452b46-2b11-4a4f-893e-22bd995151bc" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.172455 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-catalog-content\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.172940 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-utilities\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.173012 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg7bs\" (UniqueName: \"kubernetes.io/projected/03654773-2bd1-40a5-a206-82c6dfe8359f-kube-api-access-fg7bs\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.274508 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-catalog-content\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.274550 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-utilities\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.274639 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg7bs\" (UniqueName: \"kubernetes.io/projected/03654773-2bd1-40a5-a206-82c6dfe8359f-kube-api-access-fg7bs\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.275850 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-catalog-content\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.276364 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-utilities\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.296292 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg7bs\" (UniqueName: \"kubernetes.io/projected/03654773-2bd1-40a5-a206-82c6dfe8359f-kube-api-access-fg7bs\") pod \"redhat-operators-8qsds\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.369738 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.381856 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r2vnv"] Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.448499 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:18 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:18 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:18 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.448571 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.606281 4906 generic.go:334] "Generic (PLEG): container finished" podID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerID="55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc" exitCode=0 Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.606648 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxrrv" event={"ID":"0dc395cc-508a-42a1-ab5a-06cc4d3f0167","Type":"ContainerDied","Data":"55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc"} Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.608786 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r2vnv" event={"ID":"7147ecc9-6043-4838-9c47-762052b68bf1","Type":"ContainerStarted","Data":"8f33b4c49764fa55dd61c113b469240ffc2e42e9588c0b9229505e924503ad99"} Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.615430 4906 generic.go:334] "Generic (PLEG): container finished" podID="47ebbb68-ad04-4889-b010-de1719967aa7" containerID="aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c" exitCode=0 Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.616037 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-49hxc" event={"ID":"47ebbb68-ad04-4889-b010-de1719967aa7","Type":"ContainerDied","Data":"aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c"} Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.616094 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-49hxc" event={"ID":"47ebbb68-ad04-4889-b010-de1719967aa7","Type":"ContainerStarted","Data":"389038aa351f759d05f1c8930e551ef0be5337d8cb8b3026f2cfd803399bb536"} Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.633899 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8qsds"] Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.640034 4906 generic.go:334] "Generic (PLEG): container finished" podID="783a71d5-441c-49be-aad7-64d3038cac6b" containerID="8c9594c130c242c0fd021c766415c6abb1203f7ed157a2a841696c83840b1d46" exitCode=0 Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.640116 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"783a71d5-441c-49be-aad7-64d3038cac6b","Type":"ContainerDied","Data":"8c9594c130c242c0fd021c766415c6abb1203f7ed157a2a841696c83840b1d46"} Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.647800 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tc2st" Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.649619 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-7ghjq" Nov 23 06:52:18 crc kubenswrapper[4906]: W1123 06:52:18.656424 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03654773_2bd1_40a5_a206_82c6dfe8359f.slice/crio-4a340d7dbb07dee3de047e7cdbf25ca86ff41b2821f1563ad9d40d808509ec95 WatchSource:0}: Error finding container 4a340d7dbb07dee3de047e7cdbf25ca86ff41b2821f1563ad9d40d808509ec95: Status 404 returned error can't find the container with id 4a340d7dbb07dee3de047e7cdbf25ca86ff41b2821f1563ad9d40d808509ec95 Nov 23 06:52:18 crc kubenswrapper[4906]: I1123 06:52:18.991734 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.105006 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/783a71d5-441c-49be-aad7-64d3038cac6b-kube-api-access\") pod \"783a71d5-441c-49be-aad7-64d3038cac6b\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.105627 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/783a71d5-441c-49be-aad7-64d3038cac6b-kubelet-dir\") pod \"783a71d5-441c-49be-aad7-64d3038cac6b\" (UID: \"783a71d5-441c-49be-aad7-64d3038cac6b\") " Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.105792 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/783a71d5-441c-49be-aad7-64d3038cac6b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "783a71d5-441c-49be-aad7-64d3038cac6b" (UID: "783a71d5-441c-49be-aad7-64d3038cac6b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.106062 4906 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/783a71d5-441c-49be-aad7-64d3038cac6b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.112381 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/783a71d5-441c-49be-aad7-64d3038cac6b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "783a71d5-441c-49be-aad7-64d3038cac6b" (UID: "783a71d5-441c-49be-aad7-64d3038cac6b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.208061 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/783a71d5-441c-49be-aad7-64d3038cac6b-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.447775 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:19 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:19 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:19 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.447898 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.656459 4906 generic.go:334] "Generic (PLEG): container finished" podID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerID="a2969406680e2653a00727b3ef16174bc09a99b52a68f4f980f4d1ada65d2c48" exitCode=0 Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.656542 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qsds" event={"ID":"03654773-2bd1-40a5-a206-82c6dfe8359f","Type":"ContainerDied","Data":"a2969406680e2653a00727b3ef16174bc09a99b52a68f4f980f4d1ada65d2c48"} Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.656580 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qsds" event={"ID":"03654773-2bd1-40a5-a206-82c6dfe8359f","Type":"ContainerStarted","Data":"4a340d7dbb07dee3de047e7cdbf25ca86ff41b2821f1563ad9d40d808509ec95"} Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.670074 4906 generic.go:334] "Generic (PLEG): container finished" podID="7147ecc9-6043-4838-9c47-762052b68bf1" containerID="710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd" exitCode=0 Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.670237 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r2vnv" event={"ID":"7147ecc9-6043-4838-9c47-762052b68bf1","Type":"ContainerDied","Data":"710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd"} Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.684447 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"783a71d5-441c-49be-aad7-64d3038cac6b","Type":"ContainerDied","Data":"8aae5d1658eedfcb105f9b586c225acec92bb6b8f95e46d9c75b675b966f85b2"} Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.684503 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8aae5d1658eedfcb105f9b586c225acec92bb6b8f95e46d9c75b675b966f85b2" Nov 23 06:52:19 crc kubenswrapper[4906]: I1123 06:52:19.684600 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:52:20 crc kubenswrapper[4906]: I1123 06:52:20.447908 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:20 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:20 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:20 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:20 crc kubenswrapper[4906]: I1123 06:52:20.448004 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:20 crc kubenswrapper[4906]: I1123 06:52:20.945561 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:52:20 crc kubenswrapper[4906]: I1123 06:52:20.945627 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:52:21 crc kubenswrapper[4906]: I1123 06:52:21.445881 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:21 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:21 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:21 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:21 crc kubenswrapper[4906]: I1123 06:52:21.446551 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:22 crc kubenswrapper[4906]: I1123 06:52:22.447379 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:22 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:22 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:22 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:22 crc kubenswrapper[4906]: I1123 06:52:22.447490 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:22 crc kubenswrapper[4906]: I1123 06:52:22.542167 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-5txhl" Nov 23 06:52:23 crc kubenswrapper[4906]: I1123 06:52:23.445518 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:23 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:23 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:23 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:23 crc kubenswrapper[4906]: I1123 06:52:23.445616 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:24 crc kubenswrapper[4906]: I1123 06:52:24.445799 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:24 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:24 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:24 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:24 crc kubenswrapper[4906]: I1123 06:52:24.446270 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:25 crc kubenswrapper[4906]: I1123 06:52:25.445618 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:25 crc kubenswrapper[4906]: [-]has-synced failed: reason withheld Nov 23 06:52:25 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:25 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:25 crc kubenswrapper[4906]: I1123 06:52:25.445777 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:26 crc kubenswrapper[4906]: I1123 06:52:26.445419 4906 patch_prober.go:28] interesting pod/router-default-5444994796-44r49 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:52:26 crc kubenswrapper[4906]: [+]has-synced ok Nov 23 06:52:26 crc kubenswrapper[4906]: [+]process-running ok Nov 23 06:52:26 crc kubenswrapper[4906]: healthz check failed Nov 23 06:52:26 crc kubenswrapper[4906]: I1123 06:52:26.445909 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-44r49" podUID="4f13d40d-942d-4e41-99cf-3630bbc65532" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:52:26 crc kubenswrapper[4906]: I1123 06:52:26.679130 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-4qmrn" Nov 23 06:52:26 crc kubenswrapper[4906]: I1123 06:52:26.985802 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:52:26 crc kubenswrapper[4906]: I1123 06:52:26.993305 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98bb1da0-e8ce-4b13-8a4d-c30423c5baa8-metrics-certs\") pod \"network-metrics-daemon-8drn8\" (UID: \"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8\") " pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:52:27 crc kubenswrapper[4906]: I1123 06:52:27.179306 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8drn8" Nov 23 06:52:27 crc kubenswrapper[4906]: I1123 06:52:27.446029 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:27 crc kubenswrapper[4906]: I1123 06:52:27.449727 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-44r49" Nov 23 06:52:28 crc kubenswrapper[4906]: I1123 06:52:28.249373 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:28 crc kubenswrapper[4906]: I1123 06:52:28.253931 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 06:52:36 crc kubenswrapper[4906]: I1123 06:52:36.034749 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:52:40 crc kubenswrapper[4906]: E1123 06:52:40.940468 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 23 06:52:40 crc kubenswrapper[4906]: E1123 06:52:40.940989 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wjtlf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pltlf_openshift-marketplace(80bb5cda-3404-4fdc-bfb7-f13e5aa89a04): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 06:52:40 crc kubenswrapper[4906]: E1123 06:52:40.942479 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pltlf" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" Nov 23 06:52:44 crc kubenswrapper[4906]: E1123 06:52:44.763258 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-pltlf" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" Nov 23 06:52:46 crc kubenswrapper[4906]: E1123 06:52:46.222982 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 23 06:52:46 crc kubenswrapper[4906]: E1123 06:52:46.223806 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f4gkq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ln24p_openshift-marketplace(e6eaabb7-44d7-4502-958d-e2a2ed1b06ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 06:52:46 crc kubenswrapper[4906]: E1123 06:52:46.225053 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-ln24p" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" Nov 23 06:52:46 crc kubenswrapper[4906]: E1123 06:52:46.294435 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 23 06:52:46 crc kubenswrapper[4906]: E1123 06:52:46.294724 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r8pc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zsl6g_openshift-marketplace(c80737c3-e70d-40eb-a8ed-d3f4dba27233): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 06:52:46 crc kubenswrapper[4906]: E1123 06:52:46.296020 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zsl6g" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.148232 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ln24p" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.148423 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zsl6g" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.208338 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.208784 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mt64t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-49hxc_openshift-marketplace(47ebbb68-ad04-4889-b010-de1719967aa7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.210230 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-49hxc" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.420053 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.420271 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dmf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zxrrv_openshift-marketplace(0dc395cc-508a-42a1-ab5a-06cc4d3f0167): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.422660 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zxrrv" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" Nov 23 06:52:47 crc kubenswrapper[4906]: I1123 06:52:47.747319 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8drn8"] Nov 23 06:52:47 crc kubenswrapper[4906]: W1123 06:52:47.754019 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98bb1da0_e8ce_4b13_8a4d_c30423c5baa8.slice/crio-4e7555f53df8b0d23ed8b1c8aa9ba145a1c69ead555e4a8cf05a74aaec5209da WatchSource:0}: Error finding container 4e7555f53df8b0d23ed8b1c8aa9ba145a1c69ead555e4a8cf05a74aaec5209da: Status 404 returned error can't find the container with id 4e7555f53df8b0d23ed8b1c8aa9ba145a1c69ead555e4a8cf05a74aaec5209da Nov 23 06:52:47 crc kubenswrapper[4906]: I1123 06:52:47.789341 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kmzlj" Nov 23 06:52:47 crc kubenswrapper[4906]: I1123 06:52:47.896319 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8drn8" event={"ID":"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8","Type":"ContainerStarted","Data":"4e7555f53df8b0d23ed8b1c8aa9ba145a1c69ead555e4a8cf05a74aaec5209da"} Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.897044 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zxrrv" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" Nov 23 06:52:47 crc kubenswrapper[4906]: E1123 06:52:47.897248 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-49hxc" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" Nov 23 06:52:49 crc kubenswrapper[4906]: I1123 06:52:49.925500 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qsds" event={"ID":"03654773-2bd1-40a5-a206-82c6dfe8359f","Type":"ContainerStarted","Data":"dc25b32cb435389dafc502a439920a2c945cf411bafb8b3ae3f2ee9d72e7f6dd"} Nov 23 06:52:49 crc kubenswrapper[4906]: I1123 06:52:49.927765 4906 generic.go:334] "Generic (PLEG): container finished" podID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerID="10eb174a5616b1ff171ee581b0a970850961bffdf6ff655b1cae2b33a56e8a54" exitCode=0 Nov 23 06:52:49 crc kubenswrapper[4906]: I1123 06:52:49.927862 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xlcz" event={"ID":"aef8b5df-b818-47b6-9d3e-8266b55d3be7","Type":"ContainerDied","Data":"10eb174a5616b1ff171ee581b0a970850961bffdf6ff655b1cae2b33a56e8a54"} Nov 23 06:52:49 crc kubenswrapper[4906]: I1123 06:52:49.930133 4906 generic.go:334] "Generic (PLEG): container finished" podID="7147ecc9-6043-4838-9c47-762052b68bf1" containerID="2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853" exitCode=0 Nov 23 06:52:49 crc kubenswrapper[4906]: I1123 06:52:49.930191 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r2vnv" event={"ID":"7147ecc9-6043-4838-9c47-762052b68bf1","Type":"ContainerDied","Data":"2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853"} Nov 23 06:52:49 crc kubenswrapper[4906]: I1123 06:52:49.932906 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8drn8" event={"ID":"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8","Type":"ContainerStarted","Data":"be3a82309bfab52feae6df4246c2f000edf06a15fcdde1a93c88fd18c120046c"} Nov 23 06:52:49 crc kubenswrapper[4906]: I1123 06:52:49.932939 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8drn8" event={"ID":"98bb1da0-e8ce-4b13-8a4d-c30423c5baa8","Type":"ContainerStarted","Data":"66035d78ac0a43d200d172007d4cc32877fd960f4d0feb4e1d70c5efe93ac5fe"} Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.003011 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-8drn8" podStartSLOduration=166.002994202 podStartE2EDuration="2m46.002994202s" podCreationTimestamp="2025-11-23 06:50:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:49.97142784 +0000 UTC m=+185.484819143" watchObservedRunningTime="2025-11-23 06:52:50.002994202 +0000 UTC m=+185.516385505" Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.942467 4906 generic.go:334] "Generic (PLEG): container finished" podID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerID="dc25b32cb435389dafc502a439920a2c945cf411bafb8b3ae3f2ee9d72e7f6dd" exitCode=0 Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.943729 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qsds" event={"ID":"03654773-2bd1-40a5-a206-82c6dfe8359f","Type":"ContainerDied","Data":"dc25b32cb435389dafc502a439920a2c945cf411bafb8b3ae3f2ee9d72e7f6dd"} Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.945940 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.946102 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.947971 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xlcz" event={"ID":"aef8b5df-b818-47b6-9d3e-8266b55d3be7","Type":"ContainerStarted","Data":"b6a1f76a469e9307831a02aa607b926dd9eceea8de9262f3c92fd460f73f421e"} Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.950342 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r2vnv" event={"ID":"7147ecc9-6043-4838-9c47-762052b68bf1","Type":"ContainerStarted","Data":"c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b"} Nov 23 06:52:50 crc kubenswrapper[4906]: I1123 06:52:50.986927 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r2vnv" podStartSLOduration=3.346711665 podStartE2EDuration="33.986901434s" podCreationTimestamp="2025-11-23 06:52:17 +0000 UTC" firstStartedPulling="2025-11-23 06:52:19.67512737 +0000 UTC m=+155.188518673" lastFinishedPulling="2025-11-23 06:52:50.315317139 +0000 UTC m=+185.828708442" observedRunningTime="2025-11-23 06:52:50.984414879 +0000 UTC m=+186.497806192" watchObservedRunningTime="2025-11-23 06:52:50.986901434 +0000 UTC m=+186.500292737" Nov 23 06:52:51 crc kubenswrapper[4906]: I1123 06:52:51.009125 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9xlcz" podStartSLOduration=2.975846131 podStartE2EDuration="37.009096376s" podCreationTimestamp="2025-11-23 06:52:14 +0000 UTC" firstStartedPulling="2025-11-23 06:52:16.373337177 +0000 UTC m=+151.886728480" lastFinishedPulling="2025-11-23 06:52:50.406587422 +0000 UTC m=+185.919978725" observedRunningTime="2025-11-23 06:52:51.003328853 +0000 UTC m=+186.516720166" watchObservedRunningTime="2025-11-23 06:52:51.009096376 +0000 UTC m=+186.522487689" Nov 23 06:52:51 crc kubenswrapper[4906]: I1123 06:52:51.958279 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qsds" event={"ID":"03654773-2bd1-40a5-a206-82c6dfe8359f","Type":"ContainerStarted","Data":"80cb2de5b7a8cafe91ffd2759baf336df78b2814ab4e92f1ad1ac23ad1b39a46"} Nov 23 06:52:51 crc kubenswrapper[4906]: I1123 06:52:51.984760 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8qsds" podStartSLOduration=2.181709787 podStartE2EDuration="33.984737179s" podCreationTimestamp="2025-11-23 06:52:18 +0000 UTC" firstStartedPulling="2025-11-23 06:52:19.658509987 +0000 UTC m=+155.171901290" lastFinishedPulling="2025-11-23 06:52:51.461537379 +0000 UTC m=+186.974928682" observedRunningTime="2025-11-23 06:52:51.981493333 +0000 UTC m=+187.494884646" watchObservedRunningTime="2025-11-23 06:52:51.984737179 +0000 UTC m=+187.498128482" Nov 23 06:52:53 crc kubenswrapper[4906]: I1123 06:52:53.433177 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:52:55 crc kubenswrapper[4906]: I1123 06:52:55.207425 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:55 crc kubenswrapper[4906]: I1123 06:52:55.207523 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:55 crc kubenswrapper[4906]: I1123 06:52:55.644585 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:56 crc kubenswrapper[4906]: I1123 06:52:56.030418 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:56 crc kubenswrapper[4906]: I1123 06:52:56.844755 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9xlcz"] Nov 23 06:52:57 crc kubenswrapper[4906]: I1123 06:52:57.992367 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9xlcz" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="registry-server" containerID="cri-o://b6a1f76a469e9307831a02aa607b926dd9eceea8de9262f3c92fd460f73f421e" gracePeriod=2 Nov 23 06:52:58 crc kubenswrapper[4906]: I1123 06:52:58.089249 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:58 crc kubenswrapper[4906]: I1123 06:52:58.089301 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:58 crc kubenswrapper[4906]: I1123 06:52:58.130787 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:58 crc kubenswrapper[4906]: I1123 06:52:58.370774 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:58 crc kubenswrapper[4906]: I1123 06:52:58.370835 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:58 crc kubenswrapper[4906]: I1123 06:52:58.437565 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.001303 4906 generic.go:334] "Generic (PLEG): container finished" podID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerID="b6a1f76a469e9307831a02aa607b926dd9eceea8de9262f3c92fd460f73f421e" exitCode=0 Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.001383 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xlcz" event={"ID":"aef8b5df-b818-47b6-9d3e-8266b55d3be7","Type":"ContainerDied","Data":"b6a1f76a469e9307831a02aa607b926dd9eceea8de9262f3c92fd460f73f421e"} Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.078134 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.121533 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.190887 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.299776 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqfwv\" (UniqueName: \"kubernetes.io/projected/aef8b5df-b818-47b6-9d3e-8266b55d3be7-kube-api-access-kqfwv\") pod \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.299858 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-utilities\") pod \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.299891 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-catalog-content\") pod \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\" (UID: \"aef8b5df-b818-47b6-9d3e-8266b55d3be7\") " Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.301162 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-utilities" (OuterVolumeSpecName: "utilities") pod "aef8b5df-b818-47b6-9d3e-8266b55d3be7" (UID: "aef8b5df-b818-47b6-9d3e-8266b55d3be7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.306883 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef8b5df-b818-47b6-9d3e-8266b55d3be7-kube-api-access-kqfwv" (OuterVolumeSpecName: "kube-api-access-kqfwv") pod "aef8b5df-b818-47b6-9d3e-8266b55d3be7" (UID: "aef8b5df-b818-47b6-9d3e-8266b55d3be7"). InnerVolumeSpecName "kube-api-access-kqfwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.359821 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aef8b5df-b818-47b6-9d3e-8266b55d3be7" (UID: "aef8b5df-b818-47b6-9d3e-8266b55d3be7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.405400 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqfwv\" (UniqueName: \"kubernetes.io/projected/aef8b5df-b818-47b6-9d3e-8266b55d3be7-kube-api-access-kqfwv\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.405625 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:52:59 crc kubenswrapper[4906]: I1123 06:52:59.405637 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef8b5df-b818-47b6-9d3e-8266b55d3be7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.011219 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xlcz" event={"ID":"aef8b5df-b818-47b6-9d3e-8266b55d3be7","Type":"ContainerDied","Data":"bde44e6abf72230599539ea020e593d4940410a15c3cc018a9e281c3bba502c4"} Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.011318 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xlcz" Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.011321 4906 scope.go:117] "RemoveContainer" containerID="b6a1f76a469e9307831a02aa607b926dd9eceea8de9262f3c92fd460f73f421e" Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.014338 4906 generic.go:334] "Generic (PLEG): container finished" podID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerID="fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e" exitCode=0 Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.015634 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pltlf" event={"ID":"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04","Type":"ContainerDied","Data":"fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e"} Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.040495 4906 scope.go:117] "RemoveContainer" containerID="10eb174a5616b1ff171ee581b0a970850961bffdf6ff655b1cae2b33a56e8a54" Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.050242 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8qsds"] Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.053191 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9xlcz"] Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.055511 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9xlcz"] Nov 23 06:53:00 crc kubenswrapper[4906]: I1123 06:53:00.072885 4906 scope.go:117] "RemoveContainer" containerID="58822ed64ecfcd73e5382d685cf6edbaa2b6c1dcc2d7f051bed3c69b95584949" Nov 23 06:53:01 crc kubenswrapper[4906]: I1123 06:53:01.020342 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8qsds" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="registry-server" containerID="cri-o://80cb2de5b7a8cafe91ffd2759baf336df78b2814ab4e92f1ad1ac23ad1b39a46" gracePeriod=2 Nov 23 06:53:01 crc kubenswrapper[4906]: I1123 06:53:01.367423 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" path="/var/lib/kubelet/pods/aef8b5df-b818-47b6-9d3e-8266b55d3be7/volumes" Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.053301 4906 generic.go:334] "Generic (PLEG): container finished" podID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerID="80cb2de5b7a8cafe91ffd2759baf336df78b2814ab4e92f1ad1ac23ad1b39a46" exitCode=0 Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.053397 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qsds" event={"ID":"03654773-2bd1-40a5-a206-82c6dfe8359f","Type":"ContainerDied","Data":"80cb2de5b7a8cafe91ffd2759baf336df78b2814ab4e92f1ad1ac23ad1b39a46"} Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.609464 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.695234 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-utilities\") pod \"03654773-2bd1-40a5-a206-82c6dfe8359f\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.695426 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-catalog-content\") pod \"03654773-2bd1-40a5-a206-82c6dfe8359f\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.695549 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg7bs\" (UniqueName: \"kubernetes.io/projected/03654773-2bd1-40a5-a206-82c6dfe8359f-kube-api-access-fg7bs\") pod \"03654773-2bd1-40a5-a206-82c6dfe8359f\" (UID: \"03654773-2bd1-40a5-a206-82c6dfe8359f\") " Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.696628 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-utilities" (OuterVolumeSpecName: "utilities") pod "03654773-2bd1-40a5-a206-82c6dfe8359f" (UID: "03654773-2bd1-40a5-a206-82c6dfe8359f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.704941 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03654773-2bd1-40a5-a206-82c6dfe8359f-kube-api-access-fg7bs" (OuterVolumeSpecName: "kube-api-access-fg7bs") pod "03654773-2bd1-40a5-a206-82c6dfe8359f" (UID: "03654773-2bd1-40a5-a206-82c6dfe8359f"). InnerVolumeSpecName "kube-api-access-fg7bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.797205 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:03 crc kubenswrapper[4906]: I1123 06:53:03.797245 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg7bs\" (UniqueName: \"kubernetes.io/projected/03654773-2bd1-40a5-a206-82c6dfe8359f-kube-api-access-fg7bs\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.061305 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8qsds" event={"ID":"03654773-2bd1-40a5-a206-82c6dfe8359f","Type":"ContainerDied","Data":"4a340d7dbb07dee3de047e7cdbf25ca86ff41b2821f1563ad9d40d808509ec95"} Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.061377 4906 scope.go:117] "RemoveContainer" containerID="80cb2de5b7a8cafe91ffd2759baf336df78b2814ab4e92f1ad1ac23ad1b39a46" Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.061419 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8qsds" Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.117001 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03654773-2bd1-40a5-a206-82c6dfe8359f" (UID: "03654773-2bd1-40a5-a206-82c6dfe8359f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.201814 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03654773-2bd1-40a5-a206-82c6dfe8359f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.336877 4906 scope.go:117] "RemoveContainer" containerID="dc25b32cb435389dafc502a439920a2c945cf411bafb8b3ae3f2ee9d72e7f6dd" Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.378231 4906 scope.go:117] "RemoveContainer" containerID="a2969406680e2653a00727b3ef16174bc09a99b52a68f4f980f4d1ada65d2c48" Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.393616 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8qsds"] Nov 23 06:53:04 crc kubenswrapper[4906]: I1123 06:53:04.396239 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8qsds"] Nov 23 06:53:05 crc kubenswrapper[4906]: I1123 06:53:05.072349 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pltlf" event={"ID":"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04","Type":"ContainerStarted","Data":"72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796"} Nov 23 06:53:05 crc kubenswrapper[4906]: I1123 06:53:05.363829 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" path="/var/lib/kubelet/pods/03654773-2bd1-40a5-a206-82c6dfe8359f/volumes" Nov 23 06:53:06 crc kubenswrapper[4906]: I1123 06:53:06.105225 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pltlf" podStartSLOduration=4.171279921 podStartE2EDuration="52.105201968s" podCreationTimestamp="2025-11-23 06:52:14 +0000 UTC" firstStartedPulling="2025-11-23 06:52:16.402824363 +0000 UTC m=+151.916215666" lastFinishedPulling="2025-11-23 06:53:04.33674641 +0000 UTC m=+199.850137713" observedRunningTime="2025-11-23 06:53:06.102905174 +0000 UTC m=+201.616296497" watchObservedRunningTime="2025-11-23 06:53:06.105201968 +0000 UTC m=+201.618593271" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.093998 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lgph8"] Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.094622 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" podUID="aed5c517-2dbe-47af-9761-7979baf177b7" containerName="controller-manager" containerID="cri-o://59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c" gracePeriod=30 Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.120138 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerStarted","Data":"757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16"} Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.127836 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsl6g" event={"ID":"c80737c3-e70d-40eb-a8ed-d3f4dba27233","Type":"ContainerStarted","Data":"bfb4087775d230e8cf91e38603ea75f7eea436279abd80a1c5da5b0ff9d314c5"} Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.134839 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxrrv" event={"ID":"0dc395cc-508a-42a1-ab5a-06cc4d3f0167","Type":"ContainerStarted","Data":"d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca"} Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.141755 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-49hxc" event={"ID":"47ebbb68-ad04-4889-b010-de1719967aa7","Type":"ContainerStarted","Data":"a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061"} Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.204273 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t"] Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.204518 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" podUID="0b5efaa1-d4d3-4971-9afb-1490732014d6" containerName="route-controller-manager" containerID="cri-o://53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5" gracePeriod=30 Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.574584 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.579036 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647233 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed5c517-2dbe-47af-9761-7979baf177b7-serving-cert\") pod \"aed5c517-2dbe-47af-9761-7979baf177b7\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647318 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-config\") pod \"aed5c517-2dbe-47af-9761-7979baf177b7\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647397 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-proxy-ca-bundles\") pod \"aed5c517-2dbe-47af-9761-7979baf177b7\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647423 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qxmv\" (UniqueName: \"kubernetes.io/projected/aed5c517-2dbe-47af-9761-7979baf177b7-kube-api-access-8qxmv\") pod \"aed5c517-2dbe-47af-9761-7979baf177b7\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647449 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-client-ca\") pod \"aed5c517-2dbe-47af-9761-7979baf177b7\" (UID: \"aed5c517-2dbe-47af-9761-7979baf177b7\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647494 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/0b5efaa1-d4d3-4971-9afb-1490732014d6-kube-api-access-tmxtz\") pod \"0b5efaa1-d4d3-4971-9afb-1490732014d6\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647519 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-client-ca\") pod \"0b5efaa1-d4d3-4971-9afb-1490732014d6\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647558 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-config\") pod \"0b5efaa1-d4d3-4971-9afb-1490732014d6\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.647604 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5efaa1-d4d3-4971-9afb-1490732014d6-serving-cert\") pod \"0b5efaa1-d4d3-4971-9afb-1490732014d6\" (UID: \"0b5efaa1-d4d3-4971-9afb-1490732014d6\") " Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.648448 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "aed5c517-2dbe-47af-9761-7979baf177b7" (UID: "aed5c517-2dbe-47af-9761-7979baf177b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.648471 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "aed5c517-2dbe-47af-9761-7979baf177b7" (UID: "aed5c517-2dbe-47af-9761-7979baf177b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.649066 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-config" (OuterVolumeSpecName: "config") pod "aed5c517-2dbe-47af-9761-7979baf177b7" (UID: "aed5c517-2dbe-47af-9761-7979baf177b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.649167 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-config" (OuterVolumeSpecName: "config") pod "0b5efaa1-d4d3-4971-9afb-1490732014d6" (UID: "0b5efaa1-d4d3-4971-9afb-1490732014d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.649439 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-client-ca" (OuterVolumeSpecName: "client-ca") pod "0b5efaa1-d4d3-4971-9afb-1490732014d6" (UID: "0b5efaa1-d4d3-4971-9afb-1490732014d6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.654214 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b5efaa1-d4d3-4971-9afb-1490732014d6-kube-api-access-tmxtz" (OuterVolumeSpecName: "kube-api-access-tmxtz") pod "0b5efaa1-d4d3-4971-9afb-1490732014d6" (UID: "0b5efaa1-d4d3-4971-9afb-1490732014d6"). InnerVolumeSpecName "kube-api-access-tmxtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.654718 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5c517-2dbe-47af-9761-7979baf177b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "aed5c517-2dbe-47af-9761-7979baf177b7" (UID: "aed5c517-2dbe-47af-9761-7979baf177b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.654907 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed5c517-2dbe-47af-9761-7979baf177b7-kube-api-access-8qxmv" (OuterVolumeSpecName: "kube-api-access-8qxmv") pod "aed5c517-2dbe-47af-9761-7979baf177b7" (UID: "aed5c517-2dbe-47af-9761-7979baf177b7"). InnerVolumeSpecName "kube-api-access-8qxmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.657463 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b5efaa1-d4d3-4971-9afb-1490732014d6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b5efaa1-d4d3-4971-9afb-1490732014d6" (UID: "0b5efaa1-d4d3-4971-9afb-1490732014d6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749281 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed5c517-2dbe-47af-9761-7979baf177b7-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749333 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749346 4906 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749359 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qxmv\" (UniqueName: \"kubernetes.io/projected/aed5c517-2dbe-47af-9761-7979baf177b7-kube-api-access-8qxmv\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749374 4906 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed5c517-2dbe-47af-9761-7979baf177b7-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749386 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/0b5efaa1-d4d3-4971-9afb-1490732014d6-kube-api-access-tmxtz\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749396 4906 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749408 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5efaa1-d4d3-4971-9afb-1490732014d6-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:11 crc kubenswrapper[4906]: I1123 06:53:11.749418 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5efaa1-d4d3-4971-9afb-1490732014d6-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.152808 4906 generic.go:334] "Generic (PLEG): container finished" podID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerID="d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca" exitCode=0 Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.152936 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxrrv" event={"ID":"0dc395cc-508a-42a1-ab5a-06cc4d3f0167","Type":"ContainerDied","Data":"d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.157569 4906 generic.go:334] "Generic (PLEG): container finished" podID="47ebbb68-ad04-4889-b010-de1719967aa7" containerID="a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061" exitCode=0 Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.157625 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-49hxc" event={"ID":"47ebbb68-ad04-4889-b010-de1719967aa7","Type":"ContainerDied","Data":"a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.162723 4906 generic.go:334] "Generic (PLEG): container finished" podID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerID="757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16" exitCode=0 Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.162868 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerDied","Data":"757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.168043 4906 generic.go:334] "Generic (PLEG): container finished" podID="aed5c517-2dbe-47af-9761-7979baf177b7" containerID="59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c" exitCode=0 Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.168513 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.168593 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" event={"ID":"aed5c517-2dbe-47af-9761-7979baf177b7","Type":"ContainerDied","Data":"59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.168641 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lgph8" event={"ID":"aed5c517-2dbe-47af-9761-7979baf177b7","Type":"ContainerDied","Data":"0d813a908bf9f1e7c3adfa517ef48898bfdf9c6b46dcadcaa8121d597878d519"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.168672 4906 scope.go:117] "RemoveContainer" containerID="59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.172935 4906 generic.go:334] "Generic (PLEG): container finished" podID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerID="bfb4087775d230e8cf91e38603ea75f7eea436279abd80a1c5da5b0ff9d314c5" exitCode=0 Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.173034 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsl6g" event={"ID":"c80737c3-e70d-40eb-a8ed-d3f4dba27233","Type":"ContainerDied","Data":"bfb4087775d230e8cf91e38603ea75f7eea436279abd80a1c5da5b0ff9d314c5"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.175699 4906 generic.go:334] "Generic (PLEG): container finished" podID="0b5efaa1-d4d3-4971-9afb-1490732014d6" containerID="53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5" exitCode=0 Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.175770 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" event={"ID":"0b5efaa1-d4d3-4971-9afb-1490732014d6","Type":"ContainerDied","Data":"53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.175813 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" event={"ID":"0b5efaa1-d4d3-4971-9afb-1490732014d6","Type":"ContainerDied","Data":"1be19d1b1c054cf6ad8321ccb6efa014c39d4d959a24ec1496f775dd7ca6affb"} Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.175881 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.200831 4906 scope.go:117] "RemoveContainer" containerID="59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.201478 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c\": container with ID starting with 59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c not found: ID does not exist" containerID="59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.201521 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c"} err="failed to get container status \"59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c\": rpc error: code = NotFound desc = could not find container \"59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c\": container with ID starting with 59d575ba69a9afc609a8ccae0ffe3a71d76992042804640c0ca977a0dc3be93c not found: ID does not exist" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.201586 4906 scope.go:117] "RemoveContainer" containerID="53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.232326 4906 scope.go:117] "RemoveContainer" containerID="53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.233181 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5\": container with ID starting with 53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5 not found: ID does not exist" containerID="53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.233222 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5"} err="failed to get container status \"53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5\": rpc error: code = NotFound desc = could not find container \"53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5\": container with ID starting with 53037aabbc5a0ffae00570254f4a546351bcd71478e2efe97899d6f1c1daf2a5 not found: ID does not exist" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.306589 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t"] Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.316196 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-zbz6t"] Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.320852 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lgph8"] Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.323953 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lgph8"] Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.771167 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7b77cd967f-s9v9x"] Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.771790 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed5c517-2dbe-47af-9761-7979baf177b7" containerName="controller-manager" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.772289 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed5c517-2dbe-47af-9761-7979baf177b7" containerName="controller-manager" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.772821 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="registry-server" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.772865 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="registry-server" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.772889 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="extract-content" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.772899 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="extract-content" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.772950 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="extract-content" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.772960 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="extract-content" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.772986 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="registry-server" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.772995 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="registry-server" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.773502 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5efaa1-d4d3-4971-9afb-1490732014d6" containerName="route-controller-manager" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.773529 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5efaa1-d4d3-4971-9afb-1490732014d6" containerName="route-controller-manager" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.773561 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="extract-utilities" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.773572 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="extract-utilities" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.773583 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="783a71d5-441c-49be-aad7-64d3038cac6b" containerName="pruner" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.773592 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="783a71d5-441c-49be-aad7-64d3038cac6b" containerName="pruner" Nov 23 06:53:12 crc kubenswrapper[4906]: E1123 06:53:12.773613 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="extract-utilities" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.773624 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="extract-utilities" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.775365 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b5efaa1-d4d3-4971-9afb-1490732014d6" containerName="route-controller-manager" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.775396 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef8b5df-b818-47b6-9d3e-8266b55d3be7" containerName="registry-server" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.775414 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed5c517-2dbe-47af-9761-7979baf177b7" containerName="controller-manager" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.775428 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="783a71d5-441c-49be-aad7-64d3038cac6b" containerName="pruner" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.775446 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="03654773-2bd1-40a5-a206-82c6dfe8359f" containerName="registry-server" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.793738 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.795339 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8"] Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.801347 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.807290 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8"] Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.813996 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b77cd967f-s9v9x"] Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.821896 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.823266 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.823337 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.823600 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.823931 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.828222 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.828430 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.828884 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.828346 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.830860 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.830980 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.831192 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.831665 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.869985 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-config\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870040 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-client-ca\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870097 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-proxy-ca-bundles\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870148 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-config\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870377 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nnzg\" (UniqueName: \"kubernetes.io/projected/a6d4c011-c928-4da3-b293-3f353d341798-kube-api-access-4nnzg\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870458 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-client-ca\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870551 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pkdc\" (UniqueName: \"kubernetes.io/projected/d32bbeb6-888d-494c-a6bb-86757dde9aeb-kube-api-access-2pkdc\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870617 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d4c011-c928-4da3-b293-3f353d341798-serving-cert\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.870646 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32bbeb6-888d-494c-a6bb-86757dde9aeb-serving-cert\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971729 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-client-ca\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971792 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-proxy-ca-bundles\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971819 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-config\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971865 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nnzg\" (UniqueName: \"kubernetes.io/projected/a6d4c011-c928-4da3-b293-3f353d341798-kube-api-access-4nnzg\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971887 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-client-ca\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971909 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pkdc\" (UniqueName: \"kubernetes.io/projected/d32bbeb6-888d-494c-a6bb-86757dde9aeb-kube-api-access-2pkdc\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971925 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d4c011-c928-4da3-b293-3f353d341798-serving-cert\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971941 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32bbeb6-888d-494c-a6bb-86757dde9aeb-serving-cert\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.971983 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-config\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.973849 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-client-ca\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.974101 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-config\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.974240 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-client-ca\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.975284 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-config\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.975815 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-proxy-ca-bundles\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.979506 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32bbeb6-888d-494c-a6bb-86757dde9aeb-serving-cert\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:12 crc kubenswrapper[4906]: I1123 06:53:12.993604 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d4c011-c928-4da3-b293-3f353d341798-serving-cert\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.000472 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nnzg\" (UniqueName: \"kubernetes.io/projected/a6d4c011-c928-4da3-b293-3f353d341798-kube-api-access-4nnzg\") pod \"controller-manager-7b77cd967f-s9v9x\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.000714 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pkdc\" (UniqueName: \"kubernetes.io/projected/d32bbeb6-888d-494c-a6bb-86757dde9aeb-kube-api-access-2pkdc\") pod \"route-controller-manager-c779f67dc-tsmh8\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.183957 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsl6g" event={"ID":"c80737c3-e70d-40eb-a8ed-d3f4dba27233","Type":"ContainerStarted","Data":"3d3551ace3349881f441ca0d495f6dc9f1df66147d84ebfba098a010d95b279a"} Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.185987 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.186520 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxrrv" event={"ID":"0dc395cc-508a-42a1-ab5a-06cc4d3f0167","Type":"ContainerStarted","Data":"51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b"} Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.193127 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-49hxc" event={"ID":"47ebbb68-ad04-4889-b010-de1719967aa7","Type":"ContainerStarted","Data":"f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7"} Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.197893 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.202317 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerStarted","Data":"7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2"} Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.212069 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zsl6g" podStartSLOduration=2.916773253 podStartE2EDuration="59.212056042s" podCreationTimestamp="2025-11-23 06:52:14 +0000 UTC" firstStartedPulling="2025-11-23 06:52:16.388933483 +0000 UTC m=+151.902324786" lastFinishedPulling="2025-11-23 06:53:12.684216232 +0000 UTC m=+208.197607575" observedRunningTime="2025-11-23 06:53:13.207168827 +0000 UTC m=+208.720560130" watchObservedRunningTime="2025-11-23 06:53:13.212056042 +0000 UTC m=+208.725447345" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.239537 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-49hxc" podStartSLOduration=3.202948758 podStartE2EDuration="57.239508762s" podCreationTimestamp="2025-11-23 06:52:16 +0000 UTC" firstStartedPulling="2025-11-23 06:52:18.617452341 +0000 UTC m=+154.130843644" lastFinishedPulling="2025-11-23 06:53:12.654012305 +0000 UTC m=+208.167403648" observedRunningTime="2025-11-23 06:53:13.235618386 +0000 UTC m=+208.749009689" watchObservedRunningTime="2025-11-23 06:53:13.239508762 +0000 UTC m=+208.752900065" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.256747 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zxrrv" podStartSLOduration=3.240816298 podStartE2EDuration="57.256724544s" podCreationTimestamp="2025-11-23 06:52:16 +0000 UTC" firstStartedPulling="2025-11-23 06:52:18.611219324 +0000 UTC m=+154.124610617" lastFinishedPulling="2025-11-23 06:53:12.62712752 +0000 UTC m=+208.140518863" observedRunningTime="2025-11-23 06:53:13.252469177 +0000 UTC m=+208.765860490" watchObservedRunningTime="2025-11-23 06:53:13.256724544 +0000 UTC m=+208.770115847" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.288392 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ln24p" podStartSLOduration=2.953903927 podStartE2EDuration="59.288366299s" podCreationTimestamp="2025-11-23 06:52:14 +0000 UTC" firstStartedPulling="2025-11-23 06:52:16.403914182 +0000 UTC m=+151.917305485" lastFinishedPulling="2025-11-23 06:53:12.738376514 +0000 UTC m=+208.251767857" observedRunningTime="2025-11-23 06:53:13.286823386 +0000 UTC m=+208.800214699" watchObservedRunningTime="2025-11-23 06:53:13.288366299 +0000 UTC m=+208.801757602" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.383120 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b5efaa1-d4d3-4971-9afb-1490732014d6" path="/var/lib/kubelet/pods/0b5efaa1-d4d3-4971-9afb-1490732014d6/volumes" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.384011 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed5c517-2dbe-47af-9761-7979baf177b7" path="/var/lib/kubelet/pods/aed5c517-2dbe-47af-9761-7979baf177b7/volumes" Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.677641 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7b77cd967f-s9v9x"] Nov 23 06:53:13 crc kubenswrapper[4906]: I1123 06:53:13.749237 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8"] Nov 23 06:53:13 crc kubenswrapper[4906]: W1123 06:53:13.755976 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd32bbeb6_888d_494c_a6bb_86757dde9aeb.slice/crio-2a6c86af9695253d665c5bf3b165c24775fa3f46d1fe137fabc6099407a3d37f WatchSource:0}: Error finding container 2a6c86af9695253d665c5bf3b165c24775fa3f46d1fe137fabc6099407a3d37f: Status 404 returned error can't find the container with id 2a6c86af9695253d665c5bf3b165c24775fa3f46d1fe137fabc6099407a3d37f Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.208963 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" event={"ID":"d32bbeb6-888d-494c-a6bb-86757dde9aeb","Type":"ContainerStarted","Data":"8931ba2d23a9ea58b1a46db61f6d39e6ade41b49ca4f9fdfc0575bd724fc807e"} Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.209017 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" event={"ID":"d32bbeb6-888d-494c-a6bb-86757dde9aeb","Type":"ContainerStarted","Data":"2a6c86af9695253d665c5bf3b165c24775fa3f46d1fe137fabc6099407a3d37f"} Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.209330 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.211481 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" event={"ID":"a6d4c011-c928-4da3-b293-3f353d341798","Type":"ContainerStarted","Data":"35db7e6e6558be048b930f2fa04215205aa244f4d31495c3f73568eaff10be62"} Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.211514 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" event={"ID":"a6d4c011-c928-4da3-b293-3f353d341798","Type":"ContainerStarted","Data":"c03739b245cf8ea7a3cf50d71d1fbd37ba7c8327358e90bb1fab7a044f12b4c8"} Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.211732 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.226469 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.231358 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" podStartSLOduration=3.231338915 podStartE2EDuration="3.231338915s" podCreationTimestamp="2025-11-23 06:53:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:53:14.228795635 +0000 UTC m=+209.742186938" watchObservedRunningTime="2025-11-23 06:53:14.231338915 +0000 UTC m=+209.744730208" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.258100 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" podStartSLOduration=3.258074476 podStartE2EDuration="3.258074476s" podCreationTimestamp="2025-11-23 06:53:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:53:14.254892329 +0000 UTC m=+209.768283632" watchObservedRunningTime="2025-11-23 06:53:14.258074476 +0000 UTC m=+209.771465779" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.356674 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.690126 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.690401 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.817777 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.817830 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.858674 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.994166 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:53:14 crc kubenswrapper[4906]: I1123 06:53:14.995482 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:53:15 crc kubenswrapper[4906]: I1123 06:53:15.046519 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:53:15 crc kubenswrapper[4906]: I1123 06:53:15.288517 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:53:15 crc kubenswrapper[4906]: I1123 06:53:15.731601 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-ln24p" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="registry-server" probeResult="failure" output=< Nov 23 06:53:15 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 06:53:15 crc kubenswrapper[4906]: > Nov 23 06:53:16 crc kubenswrapper[4906]: I1123 06:53:16.561126 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:53:16 crc kubenswrapper[4906]: I1123 06:53:16.561186 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:53:16 crc kubenswrapper[4906]: I1123 06:53:16.612605 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:53:16 crc kubenswrapper[4906]: I1123 06:53:16.991464 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:53:16 crc kubenswrapper[4906]: I1123 06:53:16.992317 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:53:17 crc kubenswrapper[4906]: I1123 06:53:17.032278 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:53:17 crc kubenswrapper[4906]: I1123 06:53:17.270602 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:53:17 crc kubenswrapper[4906]: I1123 06:53:17.272889 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:53:18 crc kubenswrapper[4906]: I1123 06:53:18.047373 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-49hxc"] Nov 23 06:53:19 crc kubenswrapper[4906]: I1123 06:53:19.242231 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-49hxc" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="registry-server" containerID="cri-o://f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7" gracePeriod=2 Nov 23 06:53:19 crc kubenswrapper[4906]: I1123 06:53:19.787269 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:53:19 crc kubenswrapper[4906]: I1123 06:53:19.985157 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-utilities\") pod \"47ebbb68-ad04-4889-b010-de1719967aa7\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " Nov 23 06:53:19 crc kubenswrapper[4906]: I1123 06:53:19.985260 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt64t\" (UniqueName: \"kubernetes.io/projected/47ebbb68-ad04-4889-b010-de1719967aa7-kube-api-access-mt64t\") pod \"47ebbb68-ad04-4889-b010-de1719967aa7\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " Nov 23 06:53:19 crc kubenswrapper[4906]: I1123 06:53:19.985321 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-catalog-content\") pod \"47ebbb68-ad04-4889-b010-de1719967aa7\" (UID: \"47ebbb68-ad04-4889-b010-de1719967aa7\") " Nov 23 06:53:19 crc kubenswrapper[4906]: I1123 06:53:19.987333 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-utilities" (OuterVolumeSpecName: "utilities") pod "47ebbb68-ad04-4889-b010-de1719967aa7" (UID: "47ebbb68-ad04-4889-b010-de1719967aa7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:53:19 crc kubenswrapper[4906]: I1123 06:53:19.992322 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ebbb68-ad04-4889-b010-de1719967aa7-kube-api-access-mt64t" (OuterVolumeSpecName: "kube-api-access-mt64t") pod "47ebbb68-ad04-4889-b010-de1719967aa7" (UID: "47ebbb68-ad04-4889-b010-de1719967aa7"). InnerVolumeSpecName "kube-api-access-mt64t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.017672 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47ebbb68-ad04-4889-b010-de1719967aa7" (UID: "47ebbb68-ad04-4889-b010-de1719967aa7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.087106 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.087166 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt64t\" (UniqueName: \"kubernetes.io/projected/47ebbb68-ad04-4889-b010-de1719967aa7-kube-api-access-mt64t\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.087203 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebbb68-ad04-4889-b010-de1719967aa7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.254253 4906 generic.go:334] "Generic (PLEG): container finished" podID="47ebbb68-ad04-4889-b010-de1719967aa7" containerID="f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7" exitCode=0 Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.254356 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-49hxc" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.254356 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-49hxc" event={"ID":"47ebbb68-ad04-4889-b010-de1719967aa7","Type":"ContainerDied","Data":"f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7"} Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.254620 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-49hxc" event={"ID":"47ebbb68-ad04-4889-b010-de1719967aa7","Type":"ContainerDied","Data":"389038aa351f759d05f1c8930e551ef0be5337d8cb8b3026f2cfd803399bb536"} Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.254672 4906 scope.go:117] "RemoveContainer" containerID="f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.271897 4906 scope.go:117] "RemoveContainer" containerID="a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.289202 4906 scope.go:117] "RemoveContainer" containerID="aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.331989 4906 scope.go:117] "RemoveContainer" containerID="f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7" Nov 23 06:53:20 crc kubenswrapper[4906]: E1123 06:53:20.333740 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7\": container with ID starting with f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7 not found: ID does not exist" containerID="f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.333821 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7"} err="failed to get container status \"f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7\": rpc error: code = NotFound desc = could not find container \"f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7\": container with ID starting with f7eec79c40ce435608b56b0bd5c489ef00eb0e34a2ae8140eec02dc6bfc932c7 not found: ID does not exist" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.333872 4906 scope.go:117] "RemoveContainer" containerID="a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061" Nov 23 06:53:20 crc kubenswrapper[4906]: E1123 06:53:20.334226 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061\": container with ID starting with a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061 not found: ID does not exist" containerID="a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.334289 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061"} err="failed to get container status \"a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061\": rpc error: code = NotFound desc = could not find container \"a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061\": container with ID starting with a10f7cb7cb41397b3f5489a280aa759a6cf693ce86b99064a4c4a9db828b3061 not found: ID does not exist" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.334323 4906 scope.go:117] "RemoveContainer" containerID="aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c" Nov 23 06:53:20 crc kubenswrapper[4906]: E1123 06:53:20.336028 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c\": container with ID starting with aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c not found: ID does not exist" containerID="aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.336091 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c"} err="failed to get container status \"aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c\": rpc error: code = NotFound desc = could not find container \"aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c\": container with ID starting with aa15be44d69bddce4e7c9a85056b8f6fe1d8b3164f9aebcb7aa22185243fd45c not found: ID does not exist" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.339815 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-49hxc"] Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.345309 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-49hxc"] Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.946149 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.946273 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.946376 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.947848 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 06:53:20 crc kubenswrapper[4906]: I1123 06:53:20.948051 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114" gracePeriod=600 Nov 23 06:53:21 crc kubenswrapper[4906]: I1123 06:53:21.269229 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114" exitCode=0 Nov 23 06:53:21 crc kubenswrapper[4906]: I1123 06:53:21.269348 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114"} Nov 23 06:53:21 crc kubenswrapper[4906]: I1123 06:53:21.269818 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"90b994146885cb5ff4320a131a66d14febc982576aa0406885140eea76752286"} Nov 23 06:53:21 crc kubenswrapper[4906]: I1123 06:53:21.366790 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" path="/var/lib/kubelet/pods/47ebbb68-ad04-4889-b010-de1719967aa7/volumes" Nov 23 06:53:24 crc kubenswrapper[4906]: I1123 06:53:24.763904 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:53:24 crc kubenswrapper[4906]: I1123 06:53:24.837589 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:53:25 crc kubenswrapper[4906]: I1123 06:53:25.053907 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.013143 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zsl6g"] Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.013560 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zsl6g" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="registry-server" containerID="cri-o://3d3551ace3349881f441ca0d495f6dc9f1df66147d84ebfba098a010d95b279a" gracePeriod=2 Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.136801 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8hr2w"] Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.305720 4906 generic.go:334] "Generic (PLEG): container finished" podID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerID="3d3551ace3349881f441ca0d495f6dc9f1df66147d84ebfba098a010d95b279a" exitCode=0 Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.306082 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsl6g" event={"ID":"c80737c3-e70d-40eb-a8ed-d3f4dba27233","Type":"ContainerDied","Data":"3d3551ace3349881f441ca0d495f6dc9f1df66147d84ebfba098a010d95b279a"} Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.567586 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.591652 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-utilities\") pod \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.591737 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8pc5\" (UniqueName: \"kubernetes.io/projected/c80737c3-e70d-40eb-a8ed-d3f4dba27233-kube-api-access-r8pc5\") pod \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.592112 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-catalog-content\") pod \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\" (UID: \"c80737c3-e70d-40eb-a8ed-d3f4dba27233\") " Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.593820 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-utilities" (OuterVolumeSpecName: "utilities") pod "c80737c3-e70d-40eb-a8ed-d3f4dba27233" (UID: "c80737c3-e70d-40eb-a8ed-d3f4dba27233"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.613472 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c80737c3-e70d-40eb-a8ed-d3f4dba27233-kube-api-access-r8pc5" (OuterVolumeSpecName: "kube-api-access-r8pc5") pod "c80737c3-e70d-40eb-a8ed-d3f4dba27233" (UID: "c80737c3-e70d-40eb-a8ed-d3f4dba27233"). InnerVolumeSpecName "kube-api-access-r8pc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.642249 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c80737c3-e70d-40eb-a8ed-d3f4dba27233" (UID: "c80737c3-e70d-40eb-a8ed-d3f4dba27233"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.694206 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.694265 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8pc5\" (UniqueName: \"kubernetes.io/projected/c80737c3-e70d-40eb-a8ed-d3f4dba27233-kube-api-access-r8pc5\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:26 crc kubenswrapper[4906]: I1123 06:53:26.694291 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c80737c3-e70d-40eb-a8ed-d3f4dba27233-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:27 crc kubenswrapper[4906]: I1123 06:53:27.319528 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsl6g" event={"ID":"c80737c3-e70d-40eb-a8ed-d3f4dba27233","Type":"ContainerDied","Data":"e32262f90ed4869256840ba2ef6ebac59daec419d35d7ac06d42c1162c900574"} Nov 23 06:53:27 crc kubenswrapper[4906]: I1123 06:53:27.319632 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsl6g" Nov 23 06:53:27 crc kubenswrapper[4906]: I1123 06:53:27.319659 4906 scope.go:117] "RemoveContainer" containerID="3d3551ace3349881f441ca0d495f6dc9f1df66147d84ebfba098a010d95b279a" Nov 23 06:53:27 crc kubenswrapper[4906]: I1123 06:53:27.351878 4906 scope.go:117] "RemoveContainer" containerID="bfb4087775d230e8cf91e38603ea75f7eea436279abd80a1c5da5b0ff9d314c5" Nov 23 06:53:27 crc kubenswrapper[4906]: I1123 06:53:27.387322 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zsl6g"] Nov 23 06:53:27 crc kubenswrapper[4906]: I1123 06:53:27.387440 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zsl6g"] Nov 23 06:53:27 crc kubenswrapper[4906]: I1123 06:53:27.388075 4906 scope.go:117] "RemoveContainer" containerID="0d2c6987dcaa78211881fba491844ee05032a34164a9a6acaf6548507c7463f9" Nov 23 06:53:29 crc kubenswrapper[4906]: I1123 06:53:29.361901 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" path="/var/lib/kubelet/pods/c80737c3-e70d-40eb-a8ed-d3f4dba27233/volumes" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.102384 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b77cd967f-s9v9x"] Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.103007 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" podUID="a6d4c011-c928-4da3-b293-3f353d341798" containerName="controller-manager" containerID="cri-o://35db7e6e6558be048b930f2fa04215205aa244f4d31495c3f73568eaff10be62" gracePeriod=30 Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.129357 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8"] Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.130479 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" podUID="d32bbeb6-888d-494c-a6bb-86757dde9aeb" containerName="route-controller-manager" containerID="cri-o://8931ba2d23a9ea58b1a46db61f6d39e6ade41b49ca4f9fdfc0575bd724fc807e" gracePeriod=30 Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.352781 4906 generic.go:334] "Generic (PLEG): container finished" podID="a6d4c011-c928-4da3-b293-3f353d341798" containerID="35db7e6e6558be048b930f2fa04215205aa244f4d31495c3f73568eaff10be62" exitCode=0 Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.352901 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" event={"ID":"a6d4c011-c928-4da3-b293-3f353d341798","Type":"ContainerDied","Data":"35db7e6e6558be048b930f2fa04215205aa244f4d31495c3f73568eaff10be62"} Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.354424 4906 generic.go:334] "Generic (PLEG): container finished" podID="d32bbeb6-888d-494c-a6bb-86757dde9aeb" containerID="8931ba2d23a9ea58b1a46db61f6d39e6ade41b49ca4f9fdfc0575bd724fc807e" exitCode=0 Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.354452 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" event={"ID":"d32bbeb6-888d-494c-a6bb-86757dde9aeb","Type":"ContainerDied","Data":"8931ba2d23a9ea58b1a46db61f6d39e6ade41b49ca4f9fdfc0575bd724fc807e"} Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.687941 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.833172 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.874978 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-config\") pod \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.875105 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pkdc\" (UniqueName: \"kubernetes.io/projected/d32bbeb6-888d-494c-a6bb-86757dde9aeb-kube-api-access-2pkdc\") pod \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.875138 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-client-ca\") pod \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.875194 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32bbeb6-888d-494c-a6bb-86757dde9aeb-serving-cert\") pod \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\" (UID: \"d32bbeb6-888d-494c-a6bb-86757dde9aeb\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.876803 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-config" (OuterVolumeSpecName: "config") pod "d32bbeb6-888d-494c-a6bb-86757dde9aeb" (UID: "d32bbeb6-888d-494c-a6bb-86757dde9aeb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.876838 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-client-ca" (OuterVolumeSpecName: "client-ca") pod "d32bbeb6-888d-494c-a6bb-86757dde9aeb" (UID: "d32bbeb6-888d-494c-a6bb-86757dde9aeb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.883188 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32bbeb6-888d-494c-a6bb-86757dde9aeb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d32bbeb6-888d-494c-a6bb-86757dde9aeb" (UID: "d32bbeb6-888d-494c-a6bb-86757dde9aeb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.883727 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32bbeb6-888d-494c-a6bb-86757dde9aeb-kube-api-access-2pkdc" (OuterVolumeSpecName: "kube-api-access-2pkdc") pod "d32bbeb6-888d-494c-a6bb-86757dde9aeb" (UID: "d32bbeb6-888d-494c-a6bb-86757dde9aeb"). InnerVolumeSpecName "kube-api-access-2pkdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.976956 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nnzg\" (UniqueName: \"kubernetes.io/projected/a6d4c011-c928-4da3-b293-3f353d341798-kube-api-access-4nnzg\") pod \"a6d4c011-c928-4da3-b293-3f353d341798\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977068 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-config\") pod \"a6d4c011-c928-4da3-b293-3f353d341798\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977159 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-client-ca\") pod \"a6d4c011-c928-4da3-b293-3f353d341798\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977187 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d4c011-c928-4da3-b293-3f353d341798-serving-cert\") pod \"a6d4c011-c928-4da3-b293-3f353d341798\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977241 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-proxy-ca-bundles\") pod \"a6d4c011-c928-4da3-b293-3f353d341798\" (UID: \"a6d4c011-c928-4da3-b293-3f353d341798\") " Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977529 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32bbeb6-888d-494c-a6bb-86757dde9aeb-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977547 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977559 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pkdc\" (UniqueName: \"kubernetes.io/projected/d32bbeb6-888d-494c-a6bb-86757dde9aeb-kube-api-access-2pkdc\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.977572 4906 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d32bbeb6-888d-494c-a6bb-86757dde9aeb-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.978749 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a6d4c011-c928-4da3-b293-3f353d341798" (UID: "a6d4c011-c928-4da3-b293-3f353d341798"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.978823 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-client-ca" (OuterVolumeSpecName: "client-ca") pod "a6d4c011-c928-4da3-b293-3f353d341798" (UID: "a6d4c011-c928-4da3-b293-3f353d341798"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.978888 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-config" (OuterVolumeSpecName: "config") pod "a6d4c011-c928-4da3-b293-3f353d341798" (UID: "a6d4c011-c928-4da3-b293-3f353d341798"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.982035 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d4c011-c928-4da3-b293-3f353d341798-kube-api-access-4nnzg" (OuterVolumeSpecName: "kube-api-access-4nnzg") pod "a6d4c011-c928-4da3-b293-3f353d341798" (UID: "a6d4c011-c928-4da3-b293-3f353d341798"). InnerVolumeSpecName "kube-api-access-4nnzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:31 crc kubenswrapper[4906]: I1123 06:53:31.982390 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6d4c011-c928-4da3-b293-3f353d341798-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a6d4c011-c928-4da3-b293-3f353d341798" (UID: "a6d4c011-c928-4da3-b293-3f353d341798"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.078568 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nnzg\" (UniqueName: \"kubernetes.io/projected/a6d4c011-c928-4da3-b293-3f353d341798-kube-api-access-4nnzg\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.078623 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.078646 4906 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.078665 4906 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d4c011-c928-4da3-b293-3f353d341798-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.078709 4906 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a6d4c011-c928-4da3-b293-3f353d341798-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.364857 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" event={"ID":"a6d4c011-c928-4da3-b293-3f353d341798","Type":"ContainerDied","Data":"c03739b245cf8ea7a3cf50d71d1fbd37ba7c8327358e90bb1fab7a044f12b4c8"} Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.365368 4906 scope.go:117] "RemoveContainer" containerID="35db7e6e6558be048b930f2fa04215205aa244f4d31495c3f73568eaff10be62" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.364917 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7b77cd967f-s9v9x" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.370159 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" event={"ID":"d32bbeb6-888d-494c-a6bb-86757dde9aeb","Type":"ContainerDied","Data":"2a6c86af9695253d665c5bf3b165c24775fa3f46d1fe137fabc6099407a3d37f"} Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.370182 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.395007 4906 scope.go:117] "RemoveContainer" containerID="8931ba2d23a9ea58b1a46db61f6d39e6ade41b49ca4f9fdfc0575bd724fc807e" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.419878 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7b77cd967f-s9v9x"] Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.425604 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7b77cd967f-s9v9x"] Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.438747 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8"] Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.444473 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c779f67dc-tsmh8"] Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784156 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-857b5fbcd8-brv4j"] Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784542 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="extract-utilities" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784571 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="extract-utilities" Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784602 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="extract-content" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784616 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="extract-content" Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784642 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d4c011-c928-4da3-b293-3f353d341798" containerName="controller-manager" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784657 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d4c011-c928-4da3-b293-3f353d341798" containerName="controller-manager" Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784677 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32bbeb6-888d-494c-a6bb-86757dde9aeb" containerName="route-controller-manager" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784718 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32bbeb6-888d-494c-a6bb-86757dde9aeb" containerName="route-controller-manager" Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784732 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="extract-content" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784744 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="extract-content" Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784763 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="registry-server" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784776 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="registry-server" Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784798 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="extract-utilities" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784811 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="extract-utilities" Nov 23 06:53:32 crc kubenswrapper[4906]: E1123 06:53:32.784830 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="registry-server" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.784843 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="registry-server" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.785019 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c80737c3-e70d-40eb-a8ed-d3f4dba27233" containerName="registry-server" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.785044 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d4c011-c928-4da3-b293-3f353d341798" containerName="controller-manager" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.785064 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ebbb68-ad04-4889-b010-de1719967aa7" containerName="registry-server" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.785082 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32bbeb6-888d-494c-a6bb-86757dde9aeb" containerName="route-controller-manager" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.785759 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.789955 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.790248 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.790507 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.790897 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.792103 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8"] Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.793197 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.795031 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv74l\" (UniqueName: \"kubernetes.io/projected/ed31e029-fc13-46b7-92fd-b330c04fe1d2-kube-api-access-mv74l\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.795192 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-config\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.795261 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed31e029-fc13-46b7-92fd-b330c04fe1d2-serving-cert\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.795356 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.795397 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.795530 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-proxy-ca-bundles\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.795593 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-client-ca\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.796068 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.796118 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.796843 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.800955 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.801488 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-857b5fbcd8-brv4j"] Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.801503 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.801505 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.805325 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.814551 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8"] Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896133 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f9hr\" (UniqueName: \"kubernetes.io/projected/e549cb88-a124-4bc8-86e8-c1fa1026348e-kube-api-access-2f9hr\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896219 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e549cb88-a124-4bc8-86e8-c1fa1026348e-client-ca\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896256 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e549cb88-a124-4bc8-86e8-c1fa1026348e-serving-cert\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896310 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-proxy-ca-bundles\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896349 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-client-ca\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896407 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv74l\" (UniqueName: \"kubernetes.io/projected/ed31e029-fc13-46b7-92fd-b330c04fe1d2-kube-api-access-mv74l\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896476 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e549cb88-a124-4bc8-86e8-c1fa1026348e-config\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896522 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-config\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.896551 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed31e029-fc13-46b7-92fd-b330c04fe1d2-serving-cert\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.897712 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-proxy-ca-bundles\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.897937 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-client-ca\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.898951 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed31e029-fc13-46b7-92fd-b330c04fe1d2-config\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.903661 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed31e029-fc13-46b7-92fd-b330c04fe1d2-serving-cert\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.918521 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv74l\" (UniqueName: \"kubernetes.io/projected/ed31e029-fc13-46b7-92fd-b330c04fe1d2-kube-api-access-mv74l\") pod \"controller-manager-857b5fbcd8-brv4j\" (UID: \"ed31e029-fc13-46b7-92fd-b330c04fe1d2\") " pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.997321 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e549cb88-a124-4bc8-86e8-c1fa1026348e-client-ca\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.997400 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e549cb88-a124-4bc8-86e8-c1fa1026348e-serving-cert\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.997508 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e549cb88-a124-4bc8-86e8-c1fa1026348e-config\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.997626 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f9hr\" (UniqueName: \"kubernetes.io/projected/e549cb88-a124-4bc8-86e8-c1fa1026348e-kube-api-access-2f9hr\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:32 crc kubenswrapper[4906]: I1123 06:53:32.999170 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e549cb88-a124-4bc8-86e8-c1fa1026348e-client-ca\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.000647 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e549cb88-a124-4bc8-86e8-c1fa1026348e-config\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.001307 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e549cb88-a124-4bc8-86e8-c1fa1026348e-serving-cert\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.022605 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f9hr\" (UniqueName: \"kubernetes.io/projected/e549cb88-a124-4bc8-86e8-c1fa1026348e-kube-api-access-2f9hr\") pod \"route-controller-manager-64cf69f894-j25h8\" (UID: \"e549cb88-a124-4bc8-86e8-c1fa1026348e\") " pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.111918 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.127094 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.353315 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-857b5fbcd8-brv4j"] Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.365444 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d4c011-c928-4da3-b293-3f353d341798" path="/var/lib/kubelet/pods/a6d4c011-c928-4da3-b293-3f353d341798/volumes" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.366500 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d32bbeb6-888d-494c-a6bb-86757dde9aeb" path="/var/lib/kubelet/pods/d32bbeb6-888d-494c-a6bb-86757dde9aeb/volumes" Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.388586 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" event={"ID":"ed31e029-fc13-46b7-92fd-b330c04fe1d2","Type":"ContainerStarted","Data":"1b56f1265d6a5a1973786ac89266af831ad691016c563737ce159377a9a743b3"} Nov 23 06:53:33 crc kubenswrapper[4906]: I1123 06:53:33.395329 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8"] Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.396647 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" event={"ID":"ed31e029-fc13-46b7-92fd-b330c04fe1d2","Type":"ContainerStarted","Data":"b91dbf0ad680b10f19ee30803466895ce69732aa8068932d3a198c35b0663540"} Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.398175 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.401734 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" event={"ID":"e549cb88-a124-4bc8-86e8-c1fa1026348e","Type":"ContainerStarted","Data":"40fade5a1670a24a3650120a2d109a02fee26b7b07245eb98fef756a269b5ca4"} Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.401878 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" event={"ID":"e549cb88-a124-4bc8-86e8-c1fa1026348e","Type":"ContainerStarted","Data":"176884d853f48d8991ee216b25a4be78913d68a7fdd30553b4caa4f2077f6ff7"} Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.401974 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.405747 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.408759 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.416881 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-857b5fbcd8-brv4j" podStartSLOduration=3.41685697 podStartE2EDuration="3.41685697s" podCreationTimestamp="2025-11-23 06:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:53:34.414741601 +0000 UTC m=+229.928132904" watchObservedRunningTime="2025-11-23 06:53:34.41685697 +0000 UTC m=+229.930248283" Nov 23 06:53:34 crc kubenswrapper[4906]: I1123 06:53:34.435498 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64cf69f894-j25h8" podStartSLOduration=3.435476809 podStartE2EDuration="3.435476809s" podCreationTimestamp="2025-11-23 06:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:53:34.432047124 +0000 UTC m=+229.945438427" watchObservedRunningTime="2025-11-23 06:53:34.435476809 +0000 UTC m=+229.948868112" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.182335 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" podUID="803fc706-4690-447f-9aa8-bb512d9136f2" containerName="oauth-openshift" containerID="cri-o://e0652494a9087622e43571ad822143c2630cd01c43165e330a02736573aae8bb" gracePeriod=15 Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.563426 4906 generic.go:334] "Generic (PLEG): container finished" podID="803fc706-4690-447f-9aa8-bb512d9136f2" containerID="e0652494a9087622e43571ad822143c2630cd01c43165e330a02736573aae8bb" exitCode=0 Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.563509 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" event={"ID":"803fc706-4690-447f-9aa8-bb512d9136f2","Type":"ContainerDied","Data":"e0652494a9087622e43571ad822143c2630cd01c43165e330a02736573aae8bb"} Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.700840 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.738901 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx"] Nov 23 06:53:51 crc kubenswrapper[4906]: E1123 06:53:51.741707 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803fc706-4690-447f-9aa8-bb512d9136f2" containerName="oauth-openshift" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.741751 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="803fc706-4690-447f-9aa8-bb512d9136f2" containerName="oauth-openshift" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.741933 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="803fc706-4690-447f-9aa8-bb512d9136f2" containerName="oauth-openshift" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.742585 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.759917 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx"] Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.890749 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-router-certs\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.890795 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-serving-cert\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.890847 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-service-ca\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.890878 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-login\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.890901 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-trusted-ca-bundle\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.890961 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/803fc706-4690-447f-9aa8-bb512d9136f2-audit-dir\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.890996 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-audit-policies\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.891088 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-482lt\" (UniqueName: \"kubernetes.io/projected/803fc706-4690-447f-9aa8-bb512d9136f2-kube-api-access-482lt\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.891139 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-ocp-branding-template\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.891159 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-provider-selection\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.891184 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-error\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.891208 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-cliconfig\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.891225 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-idp-0-file-data\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.891254 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-session\") pod \"803fc706-4690-447f-9aa8-bb512d9136f2\" (UID: \"803fc706-4690-447f-9aa8-bb512d9136f2\") " Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892393 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-session\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892447 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-audit-policies\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892473 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892509 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-error\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892531 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-service-ca\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892561 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892583 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892612 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-login\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892644 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b5154634-83ca-4af1-9067-69fa93918e31-audit-dir\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892673 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892719 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892711 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892748 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892907 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-router-certs\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892966 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngxf6\" (UniqueName: \"kubernetes.io/projected/b5154634-83ca-4af1-9067-69fa93918e31-kube-api-access-ngxf6\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892865 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.892881 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.893127 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.893387 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/803fc706-4690-447f-9aa8-bb512d9136f2-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.893385 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.898143 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.898665 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.899557 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/803fc706-4690-447f-9aa8-bb512d9136f2-kube-api-access-482lt" (OuterVolumeSpecName: "kube-api-access-482lt") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "kube-api-access-482lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.905098 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.905125 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.906020 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.910050 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.910413 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.914958 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "803fc706-4690-447f-9aa8-bb512d9136f2" (UID: "803fc706-4690-447f-9aa8-bb512d9136f2"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995019 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b5154634-83ca-4af1-9067-69fa93918e31-audit-dir\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995081 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995110 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995138 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995171 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-router-certs\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995201 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngxf6\" (UniqueName: \"kubernetes.io/projected/b5154634-83ca-4af1-9067-69fa93918e31-kube-api-access-ngxf6\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995236 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-session\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995275 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-audit-policies\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995304 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995293 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b5154634-83ca-4af1-9067-69fa93918e31-audit-dir\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995338 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-error\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995754 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-service-ca\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995783 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995804 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995829 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-login\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995876 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995888 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995901 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995913 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995926 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995935 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995947 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995956 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995968 4906 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/803fc706-4690-447f-9aa8-bb512d9136f2-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995978 4906 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/803fc706-4690-447f-9aa8-bb512d9136f2-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995987 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-482lt\" (UniqueName: \"kubernetes.io/projected/803fc706-4690-447f-9aa8-bb512d9136f2-kube-api-access-482lt\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.995999 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.996009 4906 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/803fc706-4690-447f-9aa8-bb512d9136f2-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.997154 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.997197 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-audit-policies\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.998332 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-service-ca\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.998361 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-error\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.998506 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-router-certs\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.999634 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:51 crc kubenswrapper[4906]: I1123 06:53:51.999824 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-session\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.001057 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-login\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.001467 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.001515 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.003023 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.003704 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b5154634-83ca-4af1-9067-69fa93918e31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.026984 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngxf6\" (UniqueName: \"kubernetes.io/projected/b5154634-83ca-4af1-9067-69fa93918e31-kube-api-access-ngxf6\") pod \"oauth-openshift-7bdfd997f5-dbcpx\" (UID: \"b5154634-83ca-4af1-9067-69fa93918e31\") " pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.064173 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.555722 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx"] Nov 23 06:53:52 crc kubenswrapper[4906]: W1123 06:53:52.566476 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5154634_83ca_4af1_9067_69fa93918e31.slice/crio-efa6b8f67054c9da683c9dbbac9a331ecf04f018b0a8d8112a0e1e31f1f17839 WatchSource:0}: Error finding container efa6b8f67054c9da683c9dbbac9a331ecf04f018b0a8d8112a0e1e31f1f17839: Status 404 returned error can't find the container with id efa6b8f67054c9da683c9dbbac9a331ecf04f018b0a8d8112a0e1e31f1f17839 Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.574170 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" event={"ID":"803fc706-4690-447f-9aa8-bb512d9136f2","Type":"ContainerDied","Data":"95cd6de3ced2e6c5e0ab19496c4e5692ce224811f8e4f0359117946c96139c1e"} Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.574235 4906 scope.go:117] "RemoveContainer" containerID="e0652494a9087622e43571ad822143c2630cd01c43165e330a02736573aae8bb" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.574270 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8hr2w" Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.626126 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8hr2w"] Nov 23 06:53:52 crc kubenswrapper[4906]: I1123 06:53:52.630451 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8hr2w"] Nov 23 06:53:53 crc kubenswrapper[4906]: I1123 06:53:53.372533 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="803fc706-4690-447f-9aa8-bb512d9136f2" path="/var/lib/kubelet/pods/803fc706-4690-447f-9aa8-bb512d9136f2/volumes" Nov 23 06:53:53 crc kubenswrapper[4906]: I1123 06:53:53.585834 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" event={"ID":"b5154634-83ca-4af1-9067-69fa93918e31","Type":"ContainerStarted","Data":"17ae34fbb37cf89d0e4851ea9c817026168524aea07314c1744696d12b54e8ff"} Nov 23 06:53:53 crc kubenswrapper[4906]: I1123 06:53:53.585939 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" event={"ID":"b5154634-83ca-4af1-9067-69fa93918e31","Type":"ContainerStarted","Data":"efa6b8f67054c9da683c9dbbac9a331ecf04f018b0a8d8112a0e1e31f1f17839"} Nov 23 06:53:53 crc kubenswrapper[4906]: I1123 06:53:53.586425 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:53 crc kubenswrapper[4906]: I1123 06:53:53.595986 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" Nov 23 06:53:53 crc kubenswrapper[4906]: I1123 06:53:53.624570 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7bdfd997f5-dbcpx" podStartSLOduration=27.624537301 podStartE2EDuration="27.624537301s" podCreationTimestamp="2025-11-23 06:53:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:53:53.618924938 +0000 UTC m=+249.132316291" watchObservedRunningTime="2025-11-23 06:53:53.624537301 +0000 UTC m=+249.137928644" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.044524 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ln24p"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.045567 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ln24p" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="registry-server" containerID="cri-o://7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2" gracePeriod=30 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.060346 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pltlf"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.060737 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pltlf" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="registry-server" containerID="cri-o://72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796" gracePeriod=30 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.081546 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ktpz"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.082153 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerName="marketplace-operator" containerID="cri-o://2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875" gracePeriod=30 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.093382 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxrrv"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.093651 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zxrrv" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="registry-server" containerID="cri-o://51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b" gracePeriod=30 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.104418 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r2vnv"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.104898 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r2vnv" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="registry-server" containerID="cri-o://c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b" gracePeriod=30 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.113750 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sml78"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.114970 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.130438 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sml78"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.220964 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.221405 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdlwq\" (UniqueName: \"kubernetes.io/projected/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-kube-api-access-gdlwq\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.221447 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.322917 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdlwq\" (UniqueName: \"kubernetes.io/projected/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-kube-api-access-gdlwq\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.323010 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.323080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.325333 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.330453 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.340551 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdlwq\" (UniqueName: \"kubernetes.io/projected/8d5c5e8e-1c86-49f2-8af7-158375dc4ef4-kube-api-access-gdlwq\") pod \"marketplace-operator-79b997595-sml78\" (UID: \"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4\") " pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.502899 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.510802 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.604371 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.629491 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-utilities\") pod \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.629762 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-catalog-content\") pod \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.629806 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4gkq\" (UniqueName: \"kubernetes.io/projected/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-kube-api-access-f4gkq\") pod \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\" (UID: \"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.633093 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-utilities" (OuterVolumeSpecName: "utilities") pod "e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" (UID: "e6eaabb7-44d7-4502-958d-e2a2ed1b06ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.634864 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.642832 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-kube-api-access-f4gkq" (OuterVolumeSpecName: "kube-api-access-f4gkq") pod "e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" (UID: "e6eaabb7-44d7-4502-958d-e2a2ed1b06ea"). InnerVolumeSpecName "kube-api-access-f4gkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.649948 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.662005 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.706228 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.736742 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2xcz\" (UniqueName: \"kubernetes.io/projected/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-kube-api-access-r2xcz\") pod \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.736811 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-trusted-ca\") pod \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.737097 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjtlf\" (UniqueName: \"kubernetes.io/projected/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-kube-api-access-wjtlf\") pod \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.737127 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-operator-metrics\") pod \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\" (UID: \"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.737187 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-catalog-content\") pod \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.737268 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z8vb\" (UniqueName: \"kubernetes.io/projected/7147ecc9-6043-4838-9c47-762052b68bf1-kube-api-access-8z8vb\") pod \"7147ecc9-6043-4838-9c47-762052b68bf1\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.737342 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-utilities\") pod \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\" (UID: \"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.737482 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-utilities\") pod \"7147ecc9-6043-4838-9c47-762052b68bf1\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.737566 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-catalog-content\") pod \"7147ecc9-6043-4838-9c47-762052b68bf1\" (UID: \"7147ecc9-6043-4838-9c47-762052b68bf1\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.739856 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4gkq\" (UniqueName: \"kubernetes.io/projected/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-kube-api-access-f4gkq\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.743002 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-utilities" (OuterVolumeSpecName: "utilities") pod "7147ecc9-6043-4838-9c47-762052b68bf1" (UID: "7147ecc9-6043-4838-9c47-762052b68bf1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.743665 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" (UID: "f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.750524 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-utilities" (OuterVolumeSpecName: "utilities") pod "80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" (UID: "80bb5cda-3404-4fdc-bfb7-f13e5aa89a04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.763660 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7147ecc9-6043-4838-9c47-762052b68bf1-kube-api-access-8z8vb" (OuterVolumeSpecName: "kube-api-access-8z8vb") pod "7147ecc9-6043-4838-9c47-762052b68bf1" (UID: "7147ecc9-6043-4838-9c47-762052b68bf1"). InnerVolumeSpecName "kube-api-access-8z8vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.776034 4906 generic.go:334] "Generic (PLEG): container finished" podID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerID="2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875" exitCode=0 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.776154 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" event={"ID":"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b","Type":"ContainerDied","Data":"2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.776196 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" event={"ID":"f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b","Type":"ContainerDied","Data":"31c9724b8f16bbd939f7e535354e71d9f58d0ae97b910f72b28a615ee4a46a3c"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.776221 4906 scope.go:117] "RemoveContainer" containerID="2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.776433 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8ktpz" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.786438 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-kube-api-access-r2xcz" (OuterVolumeSpecName: "kube-api-access-r2xcz") pod "f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" (UID: "f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b"). InnerVolumeSpecName "kube-api-access-r2xcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.835387 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" (UID: "f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.836076 4906 generic.go:334] "Generic (PLEG): container finished" podID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerID="51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b" exitCode=0 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.836154 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" (UID: "e6eaabb7-44d7-4502-958d-e2a2ed1b06ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.836259 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxrrv" event={"ID":"0dc395cc-508a-42a1-ab5a-06cc4d3f0167","Type":"ContainerDied","Data":"51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.836305 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxrrv" event={"ID":"0dc395cc-508a-42a1-ab5a-06cc4d3f0167","Type":"ContainerDied","Data":"31ba81d7e03dd37cc4a8550024b4b8d5339c1181e602c736f6cea8f4fef4250d"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.838081 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxrrv" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.845634 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-catalog-content\") pod \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847032 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dmf7\" (UniqueName: \"kubernetes.io/projected/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-kube-api-access-9dmf7\") pod \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847077 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-utilities\") pod \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\" (UID: \"0dc395cc-508a-42a1-ab5a-06cc4d3f0167\") " Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847923 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2xcz\" (UniqueName: \"kubernetes.io/projected/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-kube-api-access-r2xcz\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847939 4906 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847951 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847961 4906 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847971 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z8vb\" (UniqueName: \"kubernetes.io/projected/7147ecc9-6043-4838-9c47-762052b68bf1-kube-api-access-8z8vb\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847981 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847990 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.847996 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-kube-api-access-wjtlf" (OuterVolumeSpecName: "kube-api-access-wjtlf") pod "80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" (UID: "80bb5cda-3404-4fdc-bfb7-f13e5aa89a04"). InnerVolumeSpecName "kube-api-access-wjtlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.848231 4906 scope.go:117] "RemoveContainer" containerID="2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875" Nov 23 06:54:12 crc kubenswrapper[4906]: E1123 06:54:12.849957 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875\": container with ID starting with 2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875 not found: ID does not exist" containerID="2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.850029 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875"} err="failed to get container status \"2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875\": rpc error: code = NotFound desc = could not find container \"2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875\": container with ID starting with 2850b668144f5b1923be4df59d24c104325237f457874bf4ce3612356aa86875 not found: ID does not exist" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.850078 4906 scope.go:117] "RemoveContainer" containerID="51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.850176 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-utilities" (OuterVolumeSpecName: "utilities") pod "0dc395cc-508a-42a1-ab5a-06cc4d3f0167" (UID: "0dc395cc-508a-42a1-ab5a-06cc4d3f0167"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.854536 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-kube-api-access-9dmf7" (OuterVolumeSpecName: "kube-api-access-9dmf7") pod "0dc395cc-508a-42a1-ab5a-06cc4d3f0167" (UID: "0dc395cc-508a-42a1-ab5a-06cc4d3f0167"). InnerVolumeSpecName "kube-api-access-9dmf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.858904 4906 generic.go:334] "Generic (PLEG): container finished" podID="7147ecc9-6043-4838-9c47-762052b68bf1" containerID="c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b" exitCode=0 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.858921 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r2vnv" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.858997 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r2vnv" event={"ID":"7147ecc9-6043-4838-9c47-762052b68bf1","Type":"ContainerDied","Data":"c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.860466 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r2vnv" event={"ID":"7147ecc9-6043-4838-9c47-762052b68bf1","Type":"ContainerDied","Data":"8f33b4c49764fa55dd61c113b469240ffc2e42e9588c0b9229505e924503ad99"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.870069 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dc395cc-508a-42a1-ab5a-06cc4d3f0167" (UID: "0dc395cc-508a-42a1-ab5a-06cc4d3f0167"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.876988 4906 generic.go:334] "Generic (PLEG): container finished" podID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerID="72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796" exitCode=0 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.882077 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pltlf" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.884181 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pltlf" event={"ID":"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04","Type":"ContainerDied","Data":"72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.884302 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pltlf" event={"ID":"80bb5cda-3404-4fdc-bfb7-f13e5aa89a04","Type":"ContainerDied","Data":"64b1ca4355d513f73d2bac44905aae2233586400eec33ce15dd3a8e62f57eba4"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.886128 4906 scope.go:117] "RemoveContainer" containerID="d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.887541 4906 generic.go:334] "Generic (PLEG): container finished" podID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerID="7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2" exitCode=0 Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.887610 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerDied","Data":"7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.887655 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln24p" event={"ID":"e6eaabb7-44d7-4502-958d-e2a2ed1b06ea","Type":"ContainerDied","Data":"f169430360d156e7dcf65537d7813aa7299fa0bbd7015c695c4106f6cc87a80c"} Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.887796 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ln24p" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.931718 4906 scope.go:117] "RemoveContainer" containerID="55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.949065 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ln24p"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.952336 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dmf7\" (UniqueName: \"kubernetes.io/projected/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-kube-api-access-9dmf7\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.952362 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.952373 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dc395cc-508a-42a1-ab5a-06cc4d3f0167-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.952385 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjtlf\" (UniqueName: \"kubernetes.io/projected/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-kube-api-access-wjtlf\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.957573 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ln24p"] Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.976074 4906 scope.go:117] "RemoveContainer" containerID="51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b" Nov 23 06:54:12 crc kubenswrapper[4906]: E1123 06:54:12.976876 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b\": container with ID starting with 51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b not found: ID does not exist" containerID="51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.976933 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b"} err="failed to get container status \"51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b\": rpc error: code = NotFound desc = could not find container \"51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b\": container with ID starting with 51768bdc73857fbfc6ee7adae6b7b8de2a510f3142c84500e63868a1bf6ef46b not found: ID does not exist" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.976962 4906 scope.go:117] "RemoveContainer" containerID="d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca" Nov 23 06:54:12 crc kubenswrapper[4906]: E1123 06:54:12.978895 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca\": container with ID starting with d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca not found: ID does not exist" containerID="d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.978961 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca"} err="failed to get container status \"d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca\": rpc error: code = NotFound desc = could not find container \"d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca\": container with ID starting with d28f5664bb879aa33c962308d0c9003e05530197487e78fc8dd3b3bd152eddca not found: ID does not exist" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.979019 4906 scope.go:117] "RemoveContainer" containerID="55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc" Nov 23 06:54:12 crc kubenswrapper[4906]: E1123 06:54:12.982969 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc\": container with ID starting with 55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc not found: ID does not exist" containerID="55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.983029 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc"} err="failed to get container status \"55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc\": rpc error: code = NotFound desc = could not find container \"55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc\": container with ID starting with 55f5cad51bb632093343e60ed219830a83dbb42610895d16673d49f8645bbbdc not found: ID does not exist" Nov 23 06:54:12 crc kubenswrapper[4906]: I1123 06:54:12.983066 4906 scope.go:117] "RemoveContainer" containerID="c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.004265 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" (UID: "80bb5cda-3404-4fdc-bfb7-f13e5aa89a04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.005648 4906 scope.go:117] "RemoveContainer" containerID="2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.023965 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7147ecc9-6043-4838-9c47-762052b68bf1" (UID: "7147ecc9-6043-4838-9c47-762052b68bf1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.024415 4906 scope.go:117] "RemoveContainer" containerID="710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.041526 4906 scope.go:117] "RemoveContainer" containerID="c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.042144 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b\": container with ID starting with c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b not found: ID does not exist" containerID="c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.042181 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b"} err="failed to get container status \"c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b\": rpc error: code = NotFound desc = could not find container \"c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b\": container with ID starting with c8e1bb721dedc8df9024b7351132237188f4d79fc6d20c454969a764c864379b not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.042212 4906 scope.go:117] "RemoveContainer" containerID="2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.042576 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853\": container with ID starting with 2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853 not found: ID does not exist" containerID="2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.042592 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853"} err="failed to get container status \"2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853\": rpc error: code = NotFound desc = could not find container \"2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853\": container with ID starting with 2e1b123466b8747a4a16f8b440d717e823b5436c7aaf9d76eb32697240b7c853 not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.042605 4906 scope.go:117] "RemoveContainer" containerID="710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.043721 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd\": container with ID starting with 710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd not found: ID does not exist" containerID="710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.043799 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd"} err="failed to get container status \"710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd\": rpc error: code = NotFound desc = could not find container \"710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd\": container with ID starting with 710d569aaaa7ab7d969d7e513988e10daee0d2239e95756619b4b6e878ba79bd not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.043848 4906 scope.go:117] "RemoveContainer" containerID="72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.054137 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.054166 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7147ecc9-6043-4838-9c47-762052b68bf1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.071455 4906 scope.go:117] "RemoveContainer" containerID="fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.091933 4906 scope.go:117] "RemoveContainer" containerID="f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.103258 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ktpz"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.112621 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ktpz"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.154578 4906 scope.go:117] "RemoveContainer" containerID="72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.156239 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796\": container with ID starting with 72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796 not found: ID does not exist" containerID="72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.156302 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796"} err="failed to get container status \"72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796\": rpc error: code = NotFound desc = could not find container \"72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796\": container with ID starting with 72e1e2bf6c3aa87cca1d3a2c2e60d62b61c1a25bf4d660e020cb0d92021e0796 not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.156356 4906 scope.go:117] "RemoveContainer" containerID="fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.156929 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e\": container with ID starting with fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e not found: ID does not exist" containerID="fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.156966 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e"} err="failed to get container status \"fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e\": rpc error: code = NotFound desc = could not find container \"fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e\": container with ID starting with fbc05f36a08f38809d13f7046311c5e822736e16695dbdf05623eeec267def2e not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.157053 4906 scope.go:117] "RemoveContainer" containerID="f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.159590 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675\": container with ID starting with f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675 not found: ID does not exist" containerID="f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.159659 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675"} err="failed to get container status \"f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675\": rpc error: code = NotFound desc = could not find container \"f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675\": container with ID starting with f1551cea9fad9d88051e5ffeb831a108c185f44537b0fc697ee19537cd9e1675 not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.159718 4906 scope.go:117] "RemoveContainer" containerID="7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.173870 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxrrv"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.176379 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxrrv"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.182420 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sml78"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.230245 4906 scope.go:117] "RemoveContainer" containerID="757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.250580 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pltlf"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.263211 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pltlf"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.265791 4906 scope.go:117] "RemoveContainer" containerID="8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.271137 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r2vnv"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.271220 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r2vnv"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.289618 4906 scope.go:117] "RemoveContainer" containerID="7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.290014 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2\": container with ID starting with 7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2 not found: ID does not exist" containerID="7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.290049 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2"} err="failed to get container status \"7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2\": rpc error: code = NotFound desc = could not find container \"7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2\": container with ID starting with 7a2f65f3c63d60e92fff9b0ff359c762026a550802b941cddb3c36b621c561a2 not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.290077 4906 scope.go:117] "RemoveContainer" containerID="757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.290323 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16\": container with ID starting with 757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16 not found: ID does not exist" containerID="757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.290343 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16"} err="failed to get container status \"757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16\": rpc error: code = NotFound desc = could not find container \"757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16\": container with ID starting with 757ddc4b5e4c41f1e7f52cd1c78e5cd0b9dcae04877cea50fab336f2cc4d8e16 not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.290357 4906 scope.go:117] "RemoveContainer" containerID="8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.291117 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572\": container with ID starting with 8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572 not found: ID does not exist" containerID="8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.291146 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572"} err="failed to get container status \"8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572\": rpc error: code = NotFound desc = could not find container \"8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572\": container with ID starting with 8ffd0b385ff2c48b511c232d04692d933810fc65266dbe7c8121aeeab0260572 not found: ID does not exist" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.363722 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" path="/var/lib/kubelet/pods/0dc395cc-508a-42a1-ab5a-06cc4d3f0167/volumes" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.364834 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" path="/var/lib/kubelet/pods/7147ecc9-6043-4838-9c47-762052b68bf1/volumes" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.365445 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" path="/var/lib/kubelet/pods/80bb5cda-3404-4fdc-bfb7-f13e5aa89a04/volumes" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.366163 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" path="/var/lib/kubelet/pods/e6eaabb7-44d7-4502-958d-e2a2ed1b06ea/volumes" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.367267 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" path="/var/lib/kubelet/pods/f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b/volumes" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.660770 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dmz9g"] Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661005 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661021 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661035 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661043 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661053 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661061 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661071 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661080 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661092 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661099 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661110 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerName="marketplace-operator" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661119 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerName="marketplace-operator" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661134 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661142 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661153 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661161 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661173 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661181 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661192 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661200 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="extract-utilities" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661210 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661218 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661229 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661237 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="extract-content" Nov 23 06:54:13 crc kubenswrapper[4906]: E1123 06:54:13.661252 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661259 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661372 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="7147ecc9-6043-4838-9c47-762052b68bf1" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661390 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6eaabb7-44d7-4502-958d-e2a2ed1b06ea" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661400 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dc395cc-508a-42a1-ab5a-06cc4d3f0167" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661415 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="80bb5cda-3404-4fdc-bfb7-f13e5aa89a04" containerName="registry-server" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.661432 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3c7e28f-3a68-49ba-a2c3-aa2d16aeec9b" containerName="marketplace-operator" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.662321 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.668188 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.680717 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dmz9g"] Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.771274 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcfrw\" (UniqueName: \"kubernetes.io/projected/ec2ed178-c69d-4492-9564-05c560d68db3-kube-api-access-mcfrw\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.771803 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec2ed178-c69d-4492-9564-05c560d68db3-catalog-content\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.771951 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec2ed178-c69d-4492-9564-05c560d68db3-utilities\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.873339 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcfrw\" (UniqueName: \"kubernetes.io/projected/ec2ed178-c69d-4492-9564-05c560d68db3-kube-api-access-mcfrw\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.873477 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec2ed178-c69d-4492-9564-05c560d68db3-catalog-content\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.873564 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec2ed178-c69d-4492-9564-05c560d68db3-utilities\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.874259 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec2ed178-c69d-4492-9564-05c560d68db3-catalog-content\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.874296 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec2ed178-c69d-4492-9564-05c560d68db3-utilities\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.899903 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcfrw\" (UniqueName: \"kubernetes.io/projected/ec2ed178-c69d-4492-9564-05c560d68db3-kube-api-access-mcfrw\") pod \"certified-operators-dmz9g\" (UID: \"ec2ed178-c69d-4492-9564-05c560d68db3\") " pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.900488 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sml78" event={"ID":"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4","Type":"ContainerStarted","Data":"0ad5620211367ac9ecd838d9c6605b99c03ee7804559abbcf81e3c5341eed402"} Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.900550 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sml78" event={"ID":"8d5c5e8e-1c86-49f2-8af7-158375dc4ef4","Type":"ContainerStarted","Data":"3900222cb75eda53847bf684590e5a3dec9ed3cc07d214f64c9baa2eaf636416"} Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.900848 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.906444 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sml78" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.922976 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sml78" podStartSLOduration=1.9229541239999999 podStartE2EDuration="1.922954124s" podCreationTimestamp="2025-11-23 06:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:54:13.919328035 +0000 UTC m=+269.432719348" watchObservedRunningTime="2025-11-23 06:54:13.922954124 +0000 UTC m=+269.436345437" Nov 23 06:54:13 crc kubenswrapper[4906]: I1123 06:54:13.982029 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:14 crc kubenswrapper[4906]: I1123 06:54:14.394496 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dmz9g"] Nov 23 06:54:14 crc kubenswrapper[4906]: I1123 06:54:14.909569 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec2ed178-c69d-4492-9564-05c560d68db3" containerID="f8eb96289f463c27f69c6d085eb83232b071cc5ed97d16de546afa33cea4b14c" exitCode=0 Nov 23 06:54:14 crc kubenswrapper[4906]: I1123 06:54:14.909737 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmz9g" event={"ID":"ec2ed178-c69d-4492-9564-05c560d68db3","Type":"ContainerDied","Data":"f8eb96289f463c27f69c6d085eb83232b071cc5ed97d16de546afa33cea4b14c"} Nov 23 06:54:14 crc kubenswrapper[4906]: I1123 06:54:14.909839 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmz9g" event={"ID":"ec2ed178-c69d-4492-9564-05c560d68db3","Type":"ContainerStarted","Data":"c565c028a542b8caa44522ac112413d2b7207fac11c1a4af888da28e20b1c0f0"} Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.076181 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lffck"] Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.077407 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.082939 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lffck"] Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.085724 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.196195 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-catalog-content\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.196255 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-utilities\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.196447 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnfvf\" (UniqueName: \"kubernetes.io/projected/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-kube-api-access-bnfvf\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.297474 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnfvf\" (UniqueName: \"kubernetes.io/projected/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-kube-api-access-bnfvf\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.297975 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-catalog-content\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.297998 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-utilities\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.298400 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-catalog-content\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.298521 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-utilities\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.320449 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnfvf\" (UniqueName: \"kubernetes.io/projected/ed0a90cb-fcb6-4119-b71e-dcee407aa7e0-kube-api-access-bnfvf\") pod \"redhat-marketplace-lffck\" (UID: \"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0\") " pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.403845 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.818408 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lffck"] Nov 23 06:54:15 crc kubenswrapper[4906]: W1123 06:54:15.828234 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded0a90cb_fcb6_4119_b71e_dcee407aa7e0.slice/crio-aa5ca2291ed2a23f3a70ccda133433f388c82e6ba7a03593951e93bd4a7a4e39 WatchSource:0}: Error finding container aa5ca2291ed2a23f3a70ccda133433f388c82e6ba7a03593951e93bd4a7a4e39: Status 404 returned error can't find the container with id aa5ca2291ed2a23f3a70ccda133433f388c82e6ba7a03593951e93bd4a7a4e39 Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.936451 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmz9g" event={"ID":"ec2ed178-c69d-4492-9564-05c560d68db3","Type":"ContainerStarted","Data":"7448e5be18533f7eed62aae1abcbd7d1da7a9e777dff6aba80c8407aac3b8c9d"} Nov 23 06:54:15 crc kubenswrapper[4906]: I1123 06:54:15.939263 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lffck" event={"ID":"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0","Type":"ContainerStarted","Data":"aa5ca2291ed2a23f3a70ccda133433f388c82e6ba7a03593951e93bd4a7a4e39"} Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.060607 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zspvs"] Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.061980 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.065674 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.074061 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zspvs"] Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.209877 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bc1a64-82e5-46d0-bea5-e21e40666a88-utilities\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.210331 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bln2\" (UniqueName: \"kubernetes.io/projected/08bc1a64-82e5-46d0-bea5-e21e40666a88-kube-api-access-5bln2\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.210427 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bc1a64-82e5-46d0-bea5-e21e40666a88-catalog-content\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.311583 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bc1a64-82e5-46d0-bea5-e21e40666a88-utilities\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.311647 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bln2\" (UniqueName: \"kubernetes.io/projected/08bc1a64-82e5-46d0-bea5-e21e40666a88-kube-api-access-5bln2\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.311690 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bc1a64-82e5-46d0-bea5-e21e40666a88-catalog-content\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.312203 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bc1a64-82e5-46d0-bea5-e21e40666a88-catalog-content\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.312241 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bc1a64-82e5-46d0-bea5-e21e40666a88-utilities\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.333136 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bln2\" (UniqueName: \"kubernetes.io/projected/08bc1a64-82e5-46d0-bea5-e21e40666a88-kube-api-access-5bln2\") pod \"redhat-operators-zspvs\" (UID: \"08bc1a64-82e5-46d0-bea5-e21e40666a88\") " pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.390869 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.830945 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zspvs"] Nov 23 06:54:16 crc kubenswrapper[4906]: W1123 06:54:16.841325 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08bc1a64_82e5_46d0_bea5_e21e40666a88.slice/crio-61d679f39f8fd3897d262763d179e4503c8c4174c073c6250366b049ce7011d6 WatchSource:0}: Error finding container 61d679f39f8fd3897d262763d179e4503c8c4174c073c6250366b049ce7011d6: Status 404 returned error can't find the container with id 61d679f39f8fd3897d262763d179e4503c8c4174c073c6250366b049ce7011d6 Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.946036 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec2ed178-c69d-4492-9564-05c560d68db3" containerID="7448e5be18533f7eed62aae1abcbd7d1da7a9e777dff6aba80c8407aac3b8c9d" exitCode=0 Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.946145 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmz9g" event={"ID":"ec2ed178-c69d-4492-9564-05c560d68db3","Type":"ContainerDied","Data":"7448e5be18533f7eed62aae1abcbd7d1da7a9e777dff6aba80c8407aac3b8c9d"} Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.951665 4906 generic.go:334] "Generic (PLEG): container finished" podID="ed0a90cb-fcb6-4119-b71e-dcee407aa7e0" containerID="be2299082da9178415a3141ad4fb547e19209eac1921945b9bd49de84d297375" exitCode=0 Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.951801 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lffck" event={"ID":"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0","Type":"ContainerDied","Data":"be2299082da9178415a3141ad4fb547e19209eac1921945b9bd49de84d297375"} Nov 23 06:54:16 crc kubenswrapper[4906]: I1123 06:54:16.956650 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zspvs" event={"ID":"08bc1a64-82e5-46d0-bea5-e21e40666a88","Type":"ContainerStarted","Data":"61d679f39f8fd3897d262763d179e4503c8c4174c073c6250366b049ce7011d6"} Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.463925 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jct9r"] Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.465313 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.468393 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.484203 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jct9r"] Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.535639 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-utilities\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.535713 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7rw\" (UniqueName: \"kubernetes.io/projected/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-kube-api-access-bv7rw\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.535764 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-catalog-content\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.636868 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-catalog-content\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.636958 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-utilities\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.636990 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7rw\" (UniqueName: \"kubernetes.io/projected/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-kube-api-access-bv7rw\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.638115 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-catalog-content\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.639022 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-utilities\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.667785 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7rw\" (UniqueName: \"kubernetes.io/projected/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-kube-api-access-bv7rw\") pod \"community-operators-jct9r\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.801362 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.966542 4906 generic.go:334] "Generic (PLEG): container finished" podID="ed0a90cb-fcb6-4119-b71e-dcee407aa7e0" containerID="7d9f1e8d0d61668538abadacf500612f073e0769cddb1b9294f947711f2d105c" exitCode=0 Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.966794 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lffck" event={"ID":"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0","Type":"ContainerDied","Data":"7d9f1e8d0d61668538abadacf500612f073e0769cddb1b9294f947711f2d105c"} Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.970619 4906 generic.go:334] "Generic (PLEG): container finished" podID="08bc1a64-82e5-46d0-bea5-e21e40666a88" containerID="cdc71d91169c3d84b773e6d15890ae3f28738a0b510e7ed2371f8e8dd15fa3e6" exitCode=0 Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.970756 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zspvs" event={"ID":"08bc1a64-82e5-46d0-bea5-e21e40666a88","Type":"ContainerDied","Data":"cdc71d91169c3d84b773e6d15890ae3f28738a0b510e7ed2371f8e8dd15fa3e6"} Nov 23 06:54:17 crc kubenswrapper[4906]: I1123 06:54:17.978221 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmz9g" event={"ID":"ec2ed178-c69d-4492-9564-05c560d68db3","Type":"ContainerStarted","Data":"b060f584d5bb7cc12f3558f46c19ab5c0c7a81b6013492a65090aaf02faf55b7"} Nov 23 06:54:18 crc kubenswrapper[4906]: I1123 06:54:18.037200 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dmz9g" podStartSLOduration=2.600154731 podStartE2EDuration="5.03717448s" podCreationTimestamp="2025-11-23 06:54:13 +0000 UTC" firstStartedPulling="2025-11-23 06:54:14.91288941 +0000 UTC m=+270.426280723" lastFinishedPulling="2025-11-23 06:54:17.349909159 +0000 UTC m=+272.863300472" observedRunningTime="2025-11-23 06:54:18.033382896 +0000 UTC m=+273.546774209" watchObservedRunningTime="2025-11-23 06:54:18.03717448 +0000 UTC m=+273.550565783" Nov 23 06:54:18 crc kubenswrapper[4906]: I1123 06:54:18.258282 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jct9r"] Nov 23 06:54:18 crc kubenswrapper[4906]: I1123 06:54:18.988094 4906 generic.go:334] "Generic (PLEG): container finished" podID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerID="5d3c532bd1b018440df527bdc4080286c82eee1497b9dff66a7a45f795ead44c" exitCode=0 Nov 23 06:54:18 crc kubenswrapper[4906]: I1123 06:54:18.988774 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jct9r" event={"ID":"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b","Type":"ContainerDied","Data":"5d3c532bd1b018440df527bdc4080286c82eee1497b9dff66a7a45f795ead44c"} Nov 23 06:54:18 crc kubenswrapper[4906]: I1123 06:54:18.988827 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jct9r" event={"ID":"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b","Type":"ContainerStarted","Data":"fac107a7b23cd07ea7151cff99ddabd7382710b7ef6bda53ee1ed99561ee6342"} Nov 23 06:54:18 crc kubenswrapper[4906]: I1123 06:54:18.994112 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lffck" event={"ID":"ed0a90cb-fcb6-4119-b71e-dcee407aa7e0","Type":"ContainerStarted","Data":"4136f8947fd3bd8221dc00d291c056586b3dcc1ba3bf6b27a54ed2588b6aa958"} Nov 23 06:54:19 crc kubenswrapper[4906]: I1123 06:54:19.000300 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zspvs" event={"ID":"08bc1a64-82e5-46d0-bea5-e21e40666a88","Type":"ContainerStarted","Data":"4bcb33d1287cb6aec774ad7be132ea0ffa0fecea7e8049532e039ff6b727e4dd"} Nov 23 06:54:19 crc kubenswrapper[4906]: I1123 06:54:19.065893 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lffck" podStartSLOduration=2.536786412 podStartE2EDuration="4.06586772s" podCreationTimestamp="2025-11-23 06:54:15 +0000 UTC" firstStartedPulling="2025-11-23 06:54:16.954069254 +0000 UTC m=+272.467460557" lastFinishedPulling="2025-11-23 06:54:18.483150562 +0000 UTC m=+273.996541865" observedRunningTime="2025-11-23 06:54:19.064593336 +0000 UTC m=+274.577984639" watchObservedRunningTime="2025-11-23 06:54:19.06586772 +0000 UTC m=+274.579259023" Nov 23 06:54:20 crc kubenswrapper[4906]: I1123 06:54:20.009041 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jct9r" event={"ID":"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b","Type":"ContainerStarted","Data":"735d9b427b1266406085d53a6e4b93cffeb42f5701f0bbdd9dd756d0b60ed523"} Nov 23 06:54:20 crc kubenswrapper[4906]: I1123 06:54:20.011873 4906 generic.go:334] "Generic (PLEG): container finished" podID="08bc1a64-82e5-46d0-bea5-e21e40666a88" containerID="4bcb33d1287cb6aec774ad7be132ea0ffa0fecea7e8049532e039ff6b727e4dd" exitCode=0 Nov 23 06:54:20 crc kubenswrapper[4906]: I1123 06:54:20.011928 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zspvs" event={"ID":"08bc1a64-82e5-46d0-bea5-e21e40666a88","Type":"ContainerDied","Data":"4bcb33d1287cb6aec774ad7be132ea0ffa0fecea7e8049532e039ff6b727e4dd"} Nov 23 06:54:21 crc kubenswrapper[4906]: I1123 06:54:21.025635 4906 generic.go:334] "Generic (PLEG): container finished" podID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerID="735d9b427b1266406085d53a6e4b93cffeb42f5701f0bbdd9dd756d0b60ed523" exitCode=0 Nov 23 06:54:21 crc kubenswrapper[4906]: I1123 06:54:21.025778 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jct9r" event={"ID":"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b","Type":"ContainerDied","Data":"735d9b427b1266406085d53a6e4b93cffeb42f5701f0bbdd9dd756d0b60ed523"} Nov 23 06:54:21 crc kubenswrapper[4906]: I1123 06:54:21.037395 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zspvs" event={"ID":"08bc1a64-82e5-46d0-bea5-e21e40666a88","Type":"ContainerStarted","Data":"159dc69d84e7323ae6e437466a1e2efce0b6754e17aa3283e79de39101d1b1f7"} Nov 23 06:54:21 crc kubenswrapper[4906]: I1123 06:54:21.080125 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zspvs" podStartSLOduration=2.635131527 podStartE2EDuration="5.080098706s" podCreationTimestamp="2025-11-23 06:54:16 +0000 UTC" firstStartedPulling="2025-11-23 06:54:17.974139058 +0000 UTC m=+273.487530361" lastFinishedPulling="2025-11-23 06:54:20.419106227 +0000 UTC m=+275.932497540" observedRunningTime="2025-11-23 06:54:21.077638268 +0000 UTC m=+276.591029611" watchObservedRunningTime="2025-11-23 06:54:21.080098706 +0000 UTC m=+276.593490019" Nov 23 06:54:22 crc kubenswrapper[4906]: I1123 06:54:22.046660 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jct9r" event={"ID":"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b","Type":"ContainerStarted","Data":"fd0cae8010626e20b9d43170aad30768d64b58d4a3e8e0398ac6ce023ab679da"} Nov 23 06:54:22 crc kubenswrapper[4906]: I1123 06:54:22.071582 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jct9r" podStartSLOduration=2.5633693969999998 podStartE2EDuration="5.071563773s" podCreationTimestamp="2025-11-23 06:54:17 +0000 UTC" firstStartedPulling="2025-11-23 06:54:18.991350953 +0000 UTC m=+274.504742256" lastFinishedPulling="2025-11-23 06:54:21.499545309 +0000 UTC m=+277.012936632" observedRunningTime="2025-11-23 06:54:22.070135134 +0000 UTC m=+277.583526437" watchObservedRunningTime="2025-11-23 06:54:22.071563773 +0000 UTC m=+277.584955076" Nov 23 06:54:23 crc kubenswrapper[4906]: I1123 06:54:23.982804 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:23 crc kubenswrapper[4906]: I1123 06:54:23.983381 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:24 crc kubenswrapper[4906]: I1123 06:54:24.050340 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:24 crc kubenswrapper[4906]: I1123 06:54:24.112170 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dmz9g" Nov 23 06:54:25 crc kubenswrapper[4906]: I1123 06:54:25.405339 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:25 crc kubenswrapper[4906]: I1123 06:54:25.405914 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:25 crc kubenswrapper[4906]: I1123 06:54:25.454744 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:26 crc kubenswrapper[4906]: I1123 06:54:26.124996 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lffck" Nov 23 06:54:26 crc kubenswrapper[4906]: I1123 06:54:26.391100 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:26 crc kubenswrapper[4906]: I1123 06:54:26.391227 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:27 crc kubenswrapper[4906]: I1123 06:54:27.453803 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zspvs" podUID="08bc1a64-82e5-46d0-bea5-e21e40666a88" containerName="registry-server" probeResult="failure" output=< Nov 23 06:54:27 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 06:54:27 crc kubenswrapper[4906]: > Nov 23 06:54:27 crc kubenswrapper[4906]: I1123 06:54:27.802914 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:27 crc kubenswrapper[4906]: I1123 06:54:27.803575 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:27 crc kubenswrapper[4906]: I1123 06:54:27.870557 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:28 crc kubenswrapper[4906]: I1123 06:54:28.126032 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jct9r" Nov 23 06:54:36 crc kubenswrapper[4906]: I1123 06:54:36.434715 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:54:36 crc kubenswrapper[4906]: I1123 06:54:36.478261 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zspvs" Nov 23 06:55:50 crc kubenswrapper[4906]: I1123 06:55:50.945997 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:55:50 crc kubenswrapper[4906]: I1123 06:55:50.947078 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:56:20 crc kubenswrapper[4906]: I1123 06:56:20.946259 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:56:20 crc kubenswrapper[4906]: I1123 06:56:20.947396 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:56:50 crc kubenswrapper[4906]: I1123 06:56:50.946108 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:56:50 crc kubenswrapper[4906]: I1123 06:56:50.947067 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:56:50 crc kubenswrapper[4906]: I1123 06:56:50.947129 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 06:56:50 crc kubenswrapper[4906]: I1123 06:56:50.948257 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90b994146885cb5ff4320a131a66d14febc982576aa0406885140eea76752286"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 06:56:50 crc kubenswrapper[4906]: I1123 06:56:50.948343 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://90b994146885cb5ff4320a131a66d14febc982576aa0406885140eea76752286" gracePeriod=600 Nov 23 06:56:52 crc kubenswrapper[4906]: I1123 06:56:52.045990 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="90b994146885cb5ff4320a131a66d14febc982576aa0406885140eea76752286" exitCode=0 Nov 23 06:56:52 crc kubenswrapper[4906]: I1123 06:56:52.046131 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"90b994146885cb5ff4320a131a66d14febc982576aa0406885140eea76752286"} Nov 23 06:56:52 crc kubenswrapper[4906]: I1123 06:56:52.046727 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"dd5ec0fa7f59d5550f15a2dfacce699aa4807dde1d3ef9fa91c46235d502f959"} Nov 23 06:56:52 crc kubenswrapper[4906]: I1123 06:56:52.046778 4906 scope.go:117] "RemoveContainer" containerID="75d48c952adf9dc35491b70579d2ff967b34e185a71b5fb9c65429871aa61114" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.520936 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8mszh"] Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.522430 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.585194 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8mszh"] Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.668809 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.668909 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-registry-certificates\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.668935 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.668976 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.669003 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-trusted-ca\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.669041 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-bound-sa-token\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.669070 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6wjk\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-kube-api-access-j6wjk\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.669092 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-registry-tls\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.695848 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.771488 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-bound-sa-token\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.771561 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6wjk\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-kube-api-access-j6wjk\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.771593 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-registry-tls\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.771660 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.771730 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-registry-certificates\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.771763 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.771799 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-trusted-ca\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.774751 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.774890 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-registry-certificates\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.775175 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-trusted-ca\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.783104 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.785553 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-registry-tls\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.795542 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-bound-sa-token\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.804516 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6wjk\" (UniqueName: \"kubernetes.io/projected/7ecbc60e-26d5-4011-a4b1-0caa245fee1a-kube-api-access-j6wjk\") pod \"image-registry-66df7c8f76-8mszh\" (UID: \"7ecbc60e-26d5-4011-a4b1-0caa245fee1a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:38 crc kubenswrapper[4906]: I1123 06:57:38.850927 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:39 crc kubenswrapper[4906]: I1123 06:57:39.115883 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8mszh"] Nov 23 06:57:39 crc kubenswrapper[4906]: I1123 06:57:39.392200 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" event={"ID":"7ecbc60e-26d5-4011-a4b1-0caa245fee1a","Type":"ContainerStarted","Data":"bcf045d6f9bc4b68f733606bb37f609349703a282e9d4c211ca1d76961562869"} Nov 23 06:57:39 crc kubenswrapper[4906]: I1123 06:57:39.392817 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" event={"ID":"7ecbc60e-26d5-4011-a4b1-0caa245fee1a","Type":"ContainerStarted","Data":"c9a6f8d6095b79cf5dca7bd3a531443b7622d2a56d272fcf311115e47a30dddb"} Nov 23 06:57:39 crc kubenswrapper[4906]: I1123 06:57:39.392858 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:39 crc kubenswrapper[4906]: I1123 06:57:39.424111 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" podStartSLOduration=1.424077623 podStartE2EDuration="1.424077623s" podCreationTimestamp="2025-11-23 06:57:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:57:39.423480566 +0000 UTC m=+474.936871879" watchObservedRunningTime="2025-11-23 06:57:39.424077623 +0000 UTC m=+474.937468956" Nov 23 06:57:58 crc kubenswrapper[4906]: I1123 06:57:58.859562 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8mszh" Nov 23 06:57:58 crc kubenswrapper[4906]: I1123 06:57:58.943878 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ll4c9"] Nov 23 06:58:23 crc kubenswrapper[4906]: I1123 06:58:23.995742 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" podUID="d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" containerName="registry" containerID="cri-o://ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309" gracePeriod=30 Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.419636 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.541532 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.541962 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgr25\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-kube-api-access-rgr25\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.542089 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-installation-pull-secrets\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.543754 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-bound-sa-token\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.544008 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-trusted-ca\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.544898 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.544971 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-certificates\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.545066 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-ca-trust-extracted\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.545115 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-tls\") pod \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\" (UID: \"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a\") " Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.545451 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.546804 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.551300 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.551352 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-kube-api-access-rgr25" (OuterVolumeSpecName: "kube-api-access-rgr25") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "kube-api-access-rgr25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.554327 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.556350 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.556376 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.565753 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" (UID: "d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.646599 4906 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.646640 4906 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.646654 4906 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.646668 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgr25\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-kube-api-access-rgr25\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.646702 4906 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.646716 4906 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.764769 4906 generic.go:334] "Generic (PLEG): container finished" podID="d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" containerID="ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309" exitCode=0 Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.764845 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.764834 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" event={"ID":"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a","Type":"ContainerDied","Data":"ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309"} Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.764917 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ll4c9" event={"ID":"d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a","Type":"ContainerDied","Data":"854f5dbb0197b59c1c73e548bb61162172dbf14ca782defe2e5691e71b1b742b"} Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.764956 4906 scope.go:117] "RemoveContainer" containerID="ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.791132 4906 scope.go:117] "RemoveContainer" containerID="ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309" Nov 23 06:58:24 crc kubenswrapper[4906]: E1123 06:58:24.792634 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309\": container with ID starting with ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309 not found: ID does not exist" containerID="ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.792695 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309"} err="failed to get container status \"ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309\": rpc error: code = NotFound desc = could not find container \"ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309\": container with ID starting with ab9820d764e18de3848998f6c6e81a971211d70b741cf554ea8ab613870c8309 not found: ID does not exist" Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.804887 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ll4c9"] Nov 23 06:58:24 crc kubenswrapper[4906]: I1123 06:58:24.811081 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ll4c9"] Nov 23 06:58:25 crc kubenswrapper[4906]: I1123 06:58:25.370323 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" path="/var/lib/kubelet/pods/d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a/volumes" Nov 23 06:59:20 crc kubenswrapper[4906]: I1123 06:59:20.945487 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:59:20 crc kubenswrapper[4906]: I1123 06:59:20.946466 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:59:50 crc kubenswrapper[4906]: I1123 06:59:50.946465 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:59:50 crc kubenswrapper[4906]: I1123 06:59:50.947579 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.157570 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6"] Nov 23 07:00:00 crc kubenswrapper[4906]: E1123 07:00:00.158876 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" containerName="registry" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.158903 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" containerName="registry" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.159109 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9ee8e21-2ccd-4e3f-a6e4-da4abe2ca16a" containerName="registry" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.159792 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.162667 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.163181 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.177955 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6"] Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.353815 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dfef9e-bfcd-4419-aa27-773878c02e80-secret-volume\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.354543 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dfef9e-bfcd-4419-aa27-773878c02e80-config-volume\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.354711 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8pgc\" (UniqueName: \"kubernetes.io/projected/c2dfef9e-bfcd-4419-aa27-773878c02e80-kube-api-access-x8pgc\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.455800 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dfef9e-bfcd-4419-aa27-773878c02e80-config-volume\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.455862 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8pgc\" (UniqueName: \"kubernetes.io/projected/c2dfef9e-bfcd-4419-aa27-773878c02e80-kube-api-access-x8pgc\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.455910 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dfef9e-bfcd-4419-aa27-773878c02e80-secret-volume\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.457587 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dfef9e-bfcd-4419-aa27-773878c02e80-config-volume\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.463364 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dfef9e-bfcd-4419-aa27-773878c02e80-secret-volume\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.476598 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8pgc\" (UniqueName: \"kubernetes.io/projected/c2dfef9e-bfcd-4419-aa27-773878c02e80-kube-api-access-x8pgc\") pod \"collect-profiles-29398020-szcs6\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.486412 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:00 crc kubenswrapper[4906]: I1123 07:00:00.711336 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6"] Nov 23 07:00:01 crc kubenswrapper[4906]: I1123 07:00:01.470604 4906 generic.go:334] "Generic (PLEG): container finished" podID="c2dfef9e-bfcd-4419-aa27-773878c02e80" containerID="67f58f916b7ffbc0e149b0c1d2b10f3e0edf3c34d857a2b184fc78da0a740b29" exitCode=0 Nov 23 07:00:01 crc kubenswrapper[4906]: I1123 07:00:01.470704 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" event={"ID":"c2dfef9e-bfcd-4419-aa27-773878c02e80","Type":"ContainerDied","Data":"67f58f916b7ffbc0e149b0c1d2b10f3e0edf3c34d857a2b184fc78da0a740b29"} Nov 23 07:00:01 crc kubenswrapper[4906]: I1123 07:00:01.471047 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" event={"ID":"c2dfef9e-bfcd-4419-aa27-773878c02e80","Type":"ContainerStarted","Data":"7b85c7b2db419b4a9cbe861ec056ab0c4f8b3884bb13aad179db7c6a9e4cec0a"} Nov 23 07:00:02 crc kubenswrapper[4906]: I1123 07:00:02.799653 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:02 crc kubenswrapper[4906]: I1123 07:00:02.997705 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8pgc\" (UniqueName: \"kubernetes.io/projected/c2dfef9e-bfcd-4419-aa27-773878c02e80-kube-api-access-x8pgc\") pod \"c2dfef9e-bfcd-4419-aa27-773878c02e80\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " Nov 23 07:00:02 crc kubenswrapper[4906]: I1123 07:00:02.997798 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dfef9e-bfcd-4419-aa27-773878c02e80-config-volume\") pod \"c2dfef9e-bfcd-4419-aa27-773878c02e80\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " Nov 23 07:00:02 crc kubenswrapper[4906]: I1123 07:00:02.997849 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dfef9e-bfcd-4419-aa27-773878c02e80-secret-volume\") pod \"c2dfef9e-bfcd-4419-aa27-773878c02e80\" (UID: \"c2dfef9e-bfcd-4419-aa27-773878c02e80\") " Nov 23 07:00:02 crc kubenswrapper[4906]: I1123 07:00:02.999191 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dfef9e-bfcd-4419-aa27-773878c02e80-config-volume" (OuterVolumeSpecName: "config-volume") pod "c2dfef9e-bfcd-4419-aa27-773878c02e80" (UID: "c2dfef9e-bfcd-4419-aa27-773878c02e80"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.008830 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dfef9e-bfcd-4419-aa27-773878c02e80-kube-api-access-x8pgc" (OuterVolumeSpecName: "kube-api-access-x8pgc") pod "c2dfef9e-bfcd-4419-aa27-773878c02e80" (UID: "c2dfef9e-bfcd-4419-aa27-773878c02e80"). InnerVolumeSpecName "kube-api-access-x8pgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.010802 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dfef9e-bfcd-4419-aa27-773878c02e80-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c2dfef9e-bfcd-4419-aa27-773878c02e80" (UID: "c2dfef9e-bfcd-4419-aa27-773878c02e80"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.099527 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8pgc\" (UniqueName: \"kubernetes.io/projected/c2dfef9e-bfcd-4419-aa27-773878c02e80-kube-api-access-x8pgc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.099602 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2dfef9e-bfcd-4419-aa27-773878c02e80-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.099624 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2dfef9e-bfcd-4419-aa27-773878c02e80-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.500759 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" event={"ID":"c2dfef9e-bfcd-4419-aa27-773878c02e80","Type":"ContainerDied","Data":"7b85c7b2db419b4a9cbe861ec056ab0c4f8b3884bb13aad179db7c6a9e4cec0a"} Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.500845 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b85c7b2db419b4a9cbe861ec056ab0c4f8b3884bb13aad179db7c6a9e4cec0a" Nov 23 07:00:03 crc kubenswrapper[4906]: I1123 07:00:03.500910 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6" Nov 23 07:00:20 crc kubenswrapper[4906]: I1123 07:00:20.945437 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:00:20 crc kubenswrapper[4906]: I1123 07:00:20.946310 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:00:20 crc kubenswrapper[4906]: I1123 07:00:20.946382 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:00:20 crc kubenswrapper[4906]: I1123 07:00:20.947494 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd5ec0fa7f59d5550f15a2dfacce699aa4807dde1d3ef9fa91c46235d502f959"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:00:20 crc kubenswrapper[4906]: I1123 07:00:20.947599 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://dd5ec0fa7f59d5550f15a2dfacce699aa4807dde1d3ef9fa91c46235d502f959" gracePeriod=600 Nov 23 07:00:21 crc kubenswrapper[4906]: I1123 07:00:21.636298 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="dd5ec0fa7f59d5550f15a2dfacce699aa4807dde1d3ef9fa91c46235d502f959" exitCode=0 Nov 23 07:00:21 crc kubenswrapper[4906]: I1123 07:00:21.636363 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"dd5ec0fa7f59d5550f15a2dfacce699aa4807dde1d3ef9fa91c46235d502f959"} Nov 23 07:00:21 crc kubenswrapper[4906]: I1123 07:00:21.636783 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"5e4be889a6576c0c1dc1c3392a1a758863587cc0404504906531a0c6425bf526"} Nov 23 07:00:21 crc kubenswrapper[4906]: I1123 07:00:21.636818 4906 scope.go:117] "RemoveContainer" containerID="90b994146885cb5ff4320a131a66d14febc982576aa0406885140eea76752286" Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.666501 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tkkm2"] Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.668238 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-controller" containerID="cri-o://1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" gracePeriod=30 Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.668434 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="northd" containerID="cri-o://bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" gracePeriod=30 Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.668468 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-node" containerID="cri-o://76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" gracePeriod=30 Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.668466 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" gracePeriod=30 Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.668579 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="sbdb" containerID="cri-o://35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" gracePeriod=30 Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.668444 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="nbdb" containerID="cri-o://e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" gracePeriod=30 Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.668585 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-acl-logging" containerID="cri-o://604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" gracePeriod=30 Nov 23 07:02:07 crc kubenswrapper[4906]: I1123 07:02:07.719109 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" containerID="cri-o://1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" gracePeriod=30 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.045858 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/3.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.048833 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovn-acl-logging/0.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.049383 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovn-controller/0.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.049998 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108281 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mpt2b"] Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108559 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-acl-logging" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108576 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-acl-logging" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108589 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108596 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108605 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dfef9e-bfcd-4419-aa27-773878c02e80" containerName="collect-profiles" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108613 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dfef9e-bfcd-4419-aa27-773878c02e80" containerName="collect-profiles" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108624 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108632 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108643 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108649 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108656 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="nbdb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108663 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="nbdb" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108674 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108695 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108712 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="sbdb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108719 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="sbdb" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108727 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108734 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108740 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108746 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108755 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108766 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108777 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="northd" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108784 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="northd" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108793 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-node" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108799 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-node" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.108807 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kubecfg-setup" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108814 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kubecfg-setup" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108945 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="sbdb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108958 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108969 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-node" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108986 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.108993 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-acl-logging" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109001 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="northd" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109009 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="nbdb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109018 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109027 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dfef9e-bfcd-4419-aa27-773878c02e80" containerName="collect-profiles" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109036 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovn-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109044 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109211 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.109223 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerName="ovnkube-controller" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.111944 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255397 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-openvswitch\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255544 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovn-node-metrics-cert\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255595 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255649 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-var-lib-openvswitch\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255744 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-ovn\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255625 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255792 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-netns\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255898 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255905 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255927 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255984 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-systemd\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256135 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-env-overrides\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.255942 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256182 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-kubelet\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256258 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-bin\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256273 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256328 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-script-lib\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256348 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256408 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-node-log\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256443 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-config\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256518 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-etc-openvswitch\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256611 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-netd\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256639 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-log-socket\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256766 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hvsb\" (UniqueName: \"kubernetes.io/projected/92d6f3b5-c353-4412-8f38-989b53ddef2a-kube-api-access-4hvsb\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256841 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-ovn-kubernetes\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.257429 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-systemd-units\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256516 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-node-log" (OuterVolumeSpecName: "node-log") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256733 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256818 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-log-socket" (OuterVolumeSpecName: "log-socket") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256848 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.257545 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.257566 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-slash" (OuterVolumeSpecName: "host-slash") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256909 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.257467 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-slash\") pod \"92d6f3b5-c353-4412-8f38-989b53ddef2a\" (UID: \"92d6f3b5-c353-4412-8f38-989b53ddef2a\") " Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256915 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.256905 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.257365 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.257932 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.257987 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-ovn\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258056 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-run-netns\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258108 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovnkube-script-lib\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258205 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-log-socket\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258303 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-cni-netd\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258396 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-cni-bin\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258445 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-systemd\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258482 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-slash\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258602 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258653 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovn-node-metrics-cert\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258769 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-kubelet\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258794 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-var-lib-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258815 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbhzk\" (UniqueName: \"kubernetes.io/projected/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-kube-api-access-xbhzk\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258904 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-env-overrides\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.258958 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-run-ovn-kubernetes\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259000 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-systemd-units\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259042 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-node-log\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259086 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovnkube-config\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259112 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-etc-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259321 4906 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259355 4906 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-slash\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259382 4906 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259414 4906 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259442 4906 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259468 4906 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259493 4906 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259517 4906 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259541 4906 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259565 4906 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259812 4906 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259837 4906 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-node-log\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259862 4906 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259885 4906 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259911 4906 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259935 4906 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-log-socket\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.259969 4906 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.263643 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.264083 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d6f3b5-c353-4412-8f38-989b53ddef2a-kube-api-access-4hvsb" (OuterVolumeSpecName: "kube-api-access-4hvsb") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "kube-api-access-4hvsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.280169 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "92d6f3b5-c353-4412-8f38-989b53ddef2a" (UID: "92d6f3b5-c353-4412-8f38-989b53ddef2a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.361354 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.361705 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.361777 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovn-node-metrics-cert\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362039 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-kubelet\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362142 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-var-lib-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362171 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbhzk\" (UniqueName: \"kubernetes.io/projected/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-kube-api-access-xbhzk\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362210 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-env-overrides\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362268 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-run-ovn-kubernetes\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362284 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-var-lib-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362303 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-systemd-units\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362368 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-systemd-units\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362385 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-node-log\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362429 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-node-log\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362436 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-run-ovn-kubernetes\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362412 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-kubelet\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362457 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovnkube-config\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362553 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-etc-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362706 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362744 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-ovn\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362811 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-run-netns\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362851 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovnkube-script-lib\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362873 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-log-socket\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362871 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362897 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-cni-netd\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362902 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-ovn\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362964 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-cni-bin\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362992 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-log-socket\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362966 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-etc-openvswitch\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363071 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-run-netns\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363016 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-cni-netd\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.362942 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-cni-bin\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363232 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-systemd\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363289 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-slash\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363464 4906 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/92d6f3b5-c353-4412-8f38-989b53ddef2a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363496 4906 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/92d6f3b5-c353-4412-8f38-989b53ddef2a-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363522 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hvsb\" (UniqueName: \"kubernetes.io/projected/92d6f3b5-c353-4412-8f38-989b53ddef2a-kube-api-access-4hvsb\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363764 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-run-systemd\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363837 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-env-overrides\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.363874 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-host-slash\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.364203 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovnkube-config\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.364023 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovnkube-script-lib\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.371171 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-ovn-node-metrics-cert\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.390082 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbhzk\" (UniqueName: \"kubernetes.io/projected/dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76-kube-api-access-xbhzk\") pod \"ovnkube-node-mpt2b\" (UID: \"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76\") " pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.405974 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovnkube-controller/3.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.412206 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovn-acl-logging/0.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.416547 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tkkm2_92d6f3b5-c353-4412-8f38-989b53ddef2a/ovn-controller/0.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417181 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" exitCode=0 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417229 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" exitCode=0 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417247 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" exitCode=0 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417263 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" exitCode=0 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417279 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" exitCode=0 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417293 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" exitCode=0 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417311 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" exitCode=143 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417328 4906 generic.go:334] "Generic (PLEG): container finished" podID="92d6f3b5-c353-4412-8f38-989b53ddef2a" containerID="1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" exitCode=143 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417409 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417475 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417546 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417584 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417615 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417648 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417677 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417741 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417759 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417776 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417795 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417811 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417825 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417838 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417855 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417877 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417905 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417924 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417942 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417954 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417970 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.417986 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418002 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418018 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418034 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418049 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418072 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418099 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418119 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418136 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418152 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418168 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418184 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418199 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418215 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418231 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418250 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418274 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" event={"ID":"92d6f3b5-c353-4412-8f38-989b53ddef2a","Type":"ContainerDied","Data":"22628ec75d1275fd31a92082275804ea66d5e70954359338212c45acabe91dfc"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418302 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418327 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418346 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418363 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418382 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418398 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418417 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418433 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418448 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418463 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418499 4906 scope.go:117] "RemoveContainer" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.418947 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tkkm2" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.427401 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/2.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.428655 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/1.log" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.428788 4906 generic.go:334] "Generic (PLEG): container finished" podID="4f75f165-ef56-40be-bfd4-1843de92b356" containerID="ce3c526adbece1506c82bf469d4fbe46fa7b6b3111c30f675503011841d55715" exitCode=2 Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.428852 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerDied","Data":"ce3c526adbece1506c82bf469d4fbe46fa7b6b3111c30f675503011841d55715"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.428898 4906 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28"} Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.429748 4906 scope.go:117] "RemoveContainer" containerID="ce3c526adbece1506c82bf469d4fbe46fa7b6b3111c30f675503011841d55715" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.430911 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.491735 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tkkm2"] Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.492884 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.495138 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tkkm2"] Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.524377 4906 scope.go:117] "RemoveContainer" containerID="35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.556733 4906 scope.go:117] "RemoveContainer" containerID="e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.594562 4906 scope.go:117] "RemoveContainer" containerID="bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.627090 4906 scope.go:117] "RemoveContainer" containerID="6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.652898 4906 scope.go:117] "RemoveContainer" containerID="76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.669983 4906 scope.go:117] "RemoveContainer" containerID="604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.735073 4906 scope.go:117] "RemoveContainer" containerID="1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.754804 4906 scope.go:117] "RemoveContainer" containerID="351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.773620 4906 scope.go:117] "RemoveContainer" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.774053 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": container with ID starting with 1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f not found: ID does not exist" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.774107 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} err="failed to get container status \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": rpc error: code = NotFound desc = could not find container \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": container with ID starting with 1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.774147 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.774689 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": container with ID starting with 6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9 not found: ID does not exist" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.774728 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} err="failed to get container status \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": rpc error: code = NotFound desc = could not find container \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": container with ID starting with 6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.774756 4906 scope.go:117] "RemoveContainer" containerID="35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.775338 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": container with ID starting with 35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d not found: ID does not exist" containerID="35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.775396 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} err="failed to get container status \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": rpc error: code = NotFound desc = could not find container \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": container with ID starting with 35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.775437 4906 scope.go:117] "RemoveContainer" containerID="e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.775966 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": container with ID starting with e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1 not found: ID does not exist" containerID="e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.776054 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} err="failed to get container status \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": rpc error: code = NotFound desc = could not find container \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": container with ID starting with e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.776111 4906 scope.go:117] "RemoveContainer" containerID="bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.776587 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": container with ID starting with bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1 not found: ID does not exist" containerID="bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.776618 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} err="failed to get container status \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": rpc error: code = NotFound desc = could not find container \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": container with ID starting with bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.776638 4906 scope.go:117] "RemoveContainer" containerID="6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.780872 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": container with ID starting with 6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c not found: ID does not exist" containerID="6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.780929 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} err="failed to get container status \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": rpc error: code = NotFound desc = could not find container \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": container with ID starting with 6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.780972 4906 scope.go:117] "RemoveContainer" containerID="76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.781388 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": container with ID starting with 76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb not found: ID does not exist" containerID="76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.781428 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} err="failed to get container status \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": rpc error: code = NotFound desc = could not find container \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": container with ID starting with 76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.781453 4906 scope.go:117] "RemoveContainer" containerID="604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.781857 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": container with ID starting with 604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41 not found: ID does not exist" containerID="604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.781888 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} err="failed to get container status \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": rpc error: code = NotFound desc = could not find container \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": container with ID starting with 604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.781907 4906 scope.go:117] "RemoveContainer" containerID="1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.782214 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": container with ID starting with 1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437 not found: ID does not exist" containerID="1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.782238 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} err="failed to get container status \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": rpc error: code = NotFound desc = could not find container \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": container with ID starting with 1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.782253 4906 scope.go:117] "RemoveContainer" containerID="351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0" Nov 23 07:02:08 crc kubenswrapper[4906]: E1123 07:02:08.782548 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": container with ID starting with 351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0 not found: ID does not exist" containerID="351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.782571 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} err="failed to get container status \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": rpc error: code = NotFound desc = could not find container \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": container with ID starting with 351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.782585 4906 scope.go:117] "RemoveContainer" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.783090 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} err="failed to get container status \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": rpc error: code = NotFound desc = could not find container \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": container with ID starting with 1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.783111 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.783862 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} err="failed to get container status \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": rpc error: code = NotFound desc = could not find container \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": container with ID starting with 6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.783911 4906 scope.go:117] "RemoveContainer" containerID="35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.784442 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} err="failed to get container status \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": rpc error: code = NotFound desc = could not find container \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": container with ID starting with 35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.784472 4906 scope.go:117] "RemoveContainer" containerID="e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.784884 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} err="failed to get container status \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": rpc error: code = NotFound desc = could not find container \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": container with ID starting with e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.784928 4906 scope.go:117] "RemoveContainer" containerID="bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.785401 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} err="failed to get container status \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": rpc error: code = NotFound desc = could not find container \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": container with ID starting with bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.785425 4906 scope.go:117] "RemoveContainer" containerID="6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.785905 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} err="failed to get container status \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": rpc error: code = NotFound desc = could not find container \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": container with ID starting with 6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.785953 4906 scope.go:117] "RemoveContainer" containerID="76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.786540 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} err="failed to get container status \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": rpc error: code = NotFound desc = could not find container \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": container with ID starting with 76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.786568 4906 scope.go:117] "RemoveContainer" containerID="604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.787045 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} err="failed to get container status \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": rpc error: code = NotFound desc = could not find container \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": container with ID starting with 604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.787070 4906 scope.go:117] "RemoveContainer" containerID="1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.787549 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} err="failed to get container status \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": rpc error: code = NotFound desc = could not find container \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": container with ID starting with 1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.787595 4906 scope.go:117] "RemoveContainer" containerID="351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.788057 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} err="failed to get container status \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": rpc error: code = NotFound desc = could not find container \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": container with ID starting with 351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.788091 4906 scope.go:117] "RemoveContainer" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.788573 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} err="failed to get container status \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": rpc error: code = NotFound desc = could not find container \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": container with ID starting with 1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.788598 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.789200 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} err="failed to get container status \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": rpc error: code = NotFound desc = could not find container \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": container with ID starting with 6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.789228 4906 scope.go:117] "RemoveContainer" containerID="35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.789611 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} err="failed to get container status \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": rpc error: code = NotFound desc = could not find container \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": container with ID starting with 35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.789639 4906 scope.go:117] "RemoveContainer" containerID="e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.790036 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} err="failed to get container status \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": rpc error: code = NotFound desc = could not find container \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": container with ID starting with e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.790068 4906 scope.go:117] "RemoveContainer" containerID="bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.790460 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} err="failed to get container status \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": rpc error: code = NotFound desc = could not find container \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": container with ID starting with bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.790522 4906 scope.go:117] "RemoveContainer" containerID="6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.790956 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} err="failed to get container status \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": rpc error: code = NotFound desc = could not find container \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": container with ID starting with 6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.790983 4906 scope.go:117] "RemoveContainer" containerID="76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.791281 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} err="failed to get container status \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": rpc error: code = NotFound desc = could not find container \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": container with ID starting with 76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.791309 4906 scope.go:117] "RemoveContainer" containerID="604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.792073 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} err="failed to get container status \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": rpc error: code = NotFound desc = could not find container \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": container with ID starting with 604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.792134 4906 scope.go:117] "RemoveContainer" containerID="1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.792490 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} err="failed to get container status \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": rpc error: code = NotFound desc = could not find container \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": container with ID starting with 1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.792516 4906 scope.go:117] "RemoveContainer" containerID="351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.792822 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} err="failed to get container status \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": rpc error: code = NotFound desc = could not find container \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": container with ID starting with 351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.792860 4906 scope.go:117] "RemoveContainer" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.793142 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} err="failed to get container status \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": rpc error: code = NotFound desc = could not find container \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": container with ID starting with 1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.793163 4906 scope.go:117] "RemoveContainer" containerID="6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.793418 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9"} err="failed to get container status \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": rpc error: code = NotFound desc = could not find container \"6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9\": container with ID starting with 6e59ebd2e3994a9828e698b76e394f4e96bf3d2133c435429ca8f667a55f23e9 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.793446 4906 scope.go:117] "RemoveContainer" containerID="35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.793930 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d"} err="failed to get container status \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": rpc error: code = NotFound desc = could not find container \"35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d\": container with ID starting with 35652c15139981a6477dbebd744056539d6c10aa8b717ad1f177c8f62116b07d not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.794017 4906 scope.go:117] "RemoveContainer" containerID="e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.794494 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1"} err="failed to get container status \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": rpc error: code = NotFound desc = could not find container \"e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1\": container with ID starting with e565b1b970c32a2e8430f7ad70a3551eef6236d856d60bb3828c3857d181acf1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.794519 4906 scope.go:117] "RemoveContainer" containerID="bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.794862 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1"} err="failed to get container status \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": rpc error: code = NotFound desc = could not find container \"bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1\": container with ID starting with bb6c02da81b72d7f21de15a30820b0636144beaa89b8ebc35691313f45a509f1 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.794898 4906 scope.go:117] "RemoveContainer" containerID="6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.795243 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c"} err="failed to get container status \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": rpc error: code = NotFound desc = could not find container \"6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c\": container with ID starting with 6ae9797c6db5e01420b00b8e896df4fceaf728071d08cfa35838160123deab6c not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.795270 4906 scope.go:117] "RemoveContainer" containerID="76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.795522 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb"} err="failed to get container status \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": rpc error: code = NotFound desc = could not find container \"76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb\": container with ID starting with 76e1ae87d8918d4b3d7b06f4d7633483ea2a9c1da962944197ee36896e7e14bb not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.795545 4906 scope.go:117] "RemoveContainer" containerID="604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.795759 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41"} err="failed to get container status \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": rpc error: code = NotFound desc = could not find container \"604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41\": container with ID starting with 604dd57d69edd3d582b5f1ed9698a25380eb466d94e65aad3bb757b5e2bd2c41 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.795779 4906 scope.go:117] "RemoveContainer" containerID="1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.795989 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437"} err="failed to get container status \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": rpc error: code = NotFound desc = could not find container \"1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437\": container with ID starting with 1346f289d1f508ff0af676a1f76d80af4d9854ef7415a99f9364dbb75ba88437 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.796016 4906 scope.go:117] "RemoveContainer" containerID="351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.796220 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0"} err="failed to get container status \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": rpc error: code = NotFound desc = could not find container \"351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0\": container with ID starting with 351228d8b90516266c8435d6cf518af21e767473a92b075b42e72275aebbf8e0 not found: ID does not exist" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.796243 4906 scope.go:117] "RemoveContainer" containerID="1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f" Nov 23 07:02:08 crc kubenswrapper[4906]: I1123 07:02:08.796536 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f"} err="failed to get container status \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": rpc error: code = NotFound desc = could not find container \"1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f\": container with ID starting with 1c6418a880c31f72b2832094523cf053191b3077d1cf17fc2f6f31c2f5f4483f not found: ID does not exist" Nov 23 07:02:09 crc kubenswrapper[4906]: I1123 07:02:09.371104 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d6f3b5-c353-4412-8f38-989b53ddef2a" path="/var/lib/kubelet/pods/92d6f3b5-c353-4412-8f38-989b53ddef2a/volumes" Nov 23 07:02:09 crc kubenswrapper[4906]: I1123 07:02:09.441104 4906 generic.go:334] "Generic (PLEG): container finished" podID="dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76" containerID="b5d7b2f839c8dfe79efdd111f59bb0eb4c5436fbffec89e0357d4fad1f8c968b" exitCode=0 Nov 23 07:02:09 crc kubenswrapper[4906]: I1123 07:02:09.441186 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerDied","Data":"b5d7b2f839c8dfe79efdd111f59bb0eb4c5436fbffec89e0357d4fad1f8c968b"} Nov 23 07:02:09 crc kubenswrapper[4906]: I1123 07:02:09.441211 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"9a69ef9a6ba424c1f286717da2264339fcb03c994ea069748133e48fd8566060"} Nov 23 07:02:09 crc kubenswrapper[4906]: I1123 07:02:09.449169 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/2.log" Nov 23 07:02:09 crc kubenswrapper[4906]: I1123 07:02:09.450286 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/1.log" Nov 23 07:02:09 crc kubenswrapper[4906]: I1123 07:02:09.450374 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lm9xt" event={"ID":"4f75f165-ef56-40be-bfd4-1843de92b356","Type":"ContainerStarted","Data":"5b246f0346b0cb64c31414f15050aad66c369245276cdfa06286be53e7ee9960"} Nov 23 07:02:10 crc kubenswrapper[4906]: I1123 07:02:10.462397 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"1590dbe504c42772cb3f98ba272bf9206134c1fb8c1fe11c6f978bbad8b4095e"} Nov 23 07:02:10 crc kubenswrapper[4906]: I1123 07:02:10.463527 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"b8073e4177835f481565a3189bbb12ec06bebf88935075a8702982ded4677ebe"} Nov 23 07:02:10 crc kubenswrapper[4906]: I1123 07:02:10.463560 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"b50504932116d0857ec6a2e01c681d44d218b29ee6b4f9ab0136cadf1b44acb2"} Nov 23 07:02:10 crc kubenswrapper[4906]: I1123 07:02:10.463581 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"600189acb57a95f3de54f5acfb2602673c010786e32aff1f51c2fe03058b52cc"} Nov 23 07:02:10 crc kubenswrapper[4906]: I1123 07:02:10.463601 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"03986e8c903e2153256ff08a649d05565b636d1d736dbb7034abc540c9171124"} Nov 23 07:02:11 crc kubenswrapper[4906]: I1123 07:02:11.472758 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"726c097588f4ce6f40e694f513cec4af52febf85e577a64638d96498faa68156"} Nov 23 07:02:13 crc kubenswrapper[4906]: I1123 07:02:13.493309 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"cb7d8b28f6ad6a8cccd0c8a4e104999bd7c00a925de4c10c60265a92d2803cdd"} Nov 23 07:02:15 crc kubenswrapper[4906]: I1123 07:02:15.512632 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" event={"ID":"dcfe782f-500c-4c21-b1f5-4b1b4d6b2b76","Type":"ContainerStarted","Data":"568dd55d55f03d5b45eb5a0289cd398a367d0b2abc8e0c4b02ffc1775bb7c7dd"} Nov 23 07:02:16 crc kubenswrapper[4906]: I1123 07:02:16.528209 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:16 crc kubenswrapper[4906]: I1123 07:02:16.529050 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:16 crc kubenswrapper[4906]: I1123 07:02:16.529256 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:16 crc kubenswrapper[4906]: I1123 07:02:16.565617 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" podStartSLOduration=8.565595737 podStartE2EDuration="8.565595737s" podCreationTimestamp="2025-11-23 07:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:02:16.563612385 +0000 UTC m=+752.077003688" watchObservedRunningTime="2025-11-23 07:02:16.565595737 +0000 UTC m=+752.078987040" Nov 23 07:02:16 crc kubenswrapper[4906]: I1123 07:02:16.567855 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:16 crc kubenswrapper[4906]: I1123 07:02:16.575885 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.582546 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-975zp"] Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.584322 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.609204 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-975zp"] Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.680194 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-catalog-content\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.680270 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-utilities\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.680722 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9qvj\" (UniqueName: \"kubernetes.io/projected/c45501a6-a479-43d6-ab52-805364aa1b21-kube-api-access-d9qvj\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.782382 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-catalog-content\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.783070 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-utilities\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.783308 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9qvj\" (UniqueName: \"kubernetes.io/projected/c45501a6-a479-43d6-ab52-805364aa1b21-kube-api-access-d9qvj\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.783440 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-catalog-content\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.783775 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-utilities\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.817825 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9qvj\" (UniqueName: \"kubernetes.io/projected/c45501a6-a479-43d6-ab52-805364aa1b21-kube-api-access-d9qvj\") pod \"community-operators-975zp\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: I1123 07:02:18.905168 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: E1123 07:02:18.935911 4906 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-975zp_openshift-marketplace_c45501a6-a479-43d6-ab52-805364aa1b21_0(bc937119cf9dca11279748c2d11d0fbf0602de5aa056c08a41aff9ec4cdd8ffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 07:02:18 crc kubenswrapper[4906]: E1123 07:02:18.936022 4906 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-975zp_openshift-marketplace_c45501a6-a479-43d6-ab52-805364aa1b21_0(bc937119cf9dca11279748c2d11d0fbf0602de5aa056c08a41aff9ec4cdd8ffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: E1123 07:02:18.936051 4906 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-975zp_openshift-marketplace_c45501a6-a479-43d6-ab52-805364aa1b21_0(bc937119cf9dca11279748c2d11d0fbf0602de5aa056c08a41aff9ec4cdd8ffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:18 crc kubenswrapper[4906]: E1123 07:02:18.936119 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"community-operators-975zp_openshift-marketplace(c45501a6-a479-43d6-ab52-805364aa1b21)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"community-operators-975zp_openshift-marketplace(c45501a6-a479-43d6-ab52-805364aa1b21)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-975zp_openshift-marketplace_c45501a6-a479-43d6-ab52-805364aa1b21_0(bc937119cf9dca11279748c2d11d0fbf0602de5aa056c08a41aff9ec4cdd8ffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/community-operators-975zp" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" Nov 23 07:02:19 crc kubenswrapper[4906]: I1123 07:02:19.550348 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:19 crc kubenswrapper[4906]: I1123 07:02:19.551185 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:19 crc kubenswrapper[4906]: I1123 07:02:19.857129 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-975zp"] Nov 23 07:02:19 crc kubenswrapper[4906]: W1123 07:02:19.864194 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc45501a6_a479_43d6_ab52_805364aa1b21.slice/crio-5e9f813f2616677f0e726e7fa4ce720df7c0de2a95f2f1375189481ae6e6310a WatchSource:0}: Error finding container 5e9f813f2616677f0e726e7fa4ce720df7c0de2a95f2f1375189481ae6e6310a: Status 404 returned error can't find the container with id 5e9f813f2616677f0e726e7fa4ce720df7c0de2a95f2f1375189481ae6e6310a Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.101094 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-k7tpf"] Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.102051 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.106202 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.106455 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.107101 4906 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5h7ln" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.109393 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.115086 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-k7tpf"] Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.204056 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/10146b05-1750-4ed1-ba76-ba7899743235-crc-storage\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.204229 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjtnw\" (UniqueName: \"kubernetes.io/projected/10146b05-1750-4ed1-ba76-ba7899743235-kube-api-access-xjtnw\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.204283 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/10146b05-1750-4ed1-ba76-ba7899743235-node-mnt\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.305878 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjtnw\" (UniqueName: \"kubernetes.io/projected/10146b05-1750-4ed1-ba76-ba7899743235-kube-api-access-xjtnw\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.306006 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/10146b05-1750-4ed1-ba76-ba7899743235-node-mnt\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.306123 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/10146b05-1750-4ed1-ba76-ba7899743235-crc-storage\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.306421 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/10146b05-1750-4ed1-ba76-ba7899743235-node-mnt\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.309478 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/10146b05-1750-4ed1-ba76-ba7899743235-crc-storage\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.343988 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjtnw\" (UniqueName: \"kubernetes.io/projected/10146b05-1750-4ed1-ba76-ba7899743235-kube-api-access-xjtnw\") pod \"crc-storage-crc-k7tpf\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.485434 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.563828 4906 generic.go:334] "Generic (PLEG): container finished" podID="c45501a6-a479-43d6-ab52-805364aa1b21" containerID="65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9" exitCode=0 Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.563938 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-975zp" event={"ID":"c45501a6-a479-43d6-ab52-805364aa1b21","Type":"ContainerDied","Data":"65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9"} Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.564556 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-975zp" event={"ID":"c45501a6-a479-43d6-ab52-805364aa1b21","Type":"ContainerStarted","Data":"5e9f813f2616677f0e726e7fa4ce720df7c0de2a95f2f1375189481ae6e6310a"} Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.565470 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:02:20 crc kubenswrapper[4906]: I1123 07:02:20.798261 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-k7tpf"] Nov 23 07:02:20 crc kubenswrapper[4906]: W1123 07:02:20.811560 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10146b05_1750_4ed1_ba76_ba7899743235.slice/crio-db0555e00f4ae1894b519a0c7e60ebfaa5112a9c9472264ab4fe76826a01a50c WatchSource:0}: Error finding container db0555e00f4ae1894b519a0c7e60ebfaa5112a9c9472264ab4fe76826a01a50c: Status 404 returned error can't find the container with id db0555e00f4ae1894b519a0c7e60ebfaa5112a9c9472264ab4fe76826a01a50c Nov 23 07:02:21 crc kubenswrapper[4906]: I1123 07:02:21.239152 4906 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 07:02:21 crc kubenswrapper[4906]: I1123 07:02:21.571008 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-k7tpf" event={"ID":"10146b05-1750-4ed1-ba76-ba7899743235","Type":"ContainerStarted","Data":"db0555e00f4ae1894b519a0c7e60ebfaa5112a9c9472264ab4fe76826a01a50c"} Nov 23 07:02:22 crc kubenswrapper[4906]: I1123 07:02:22.581068 4906 generic.go:334] "Generic (PLEG): container finished" podID="10146b05-1750-4ed1-ba76-ba7899743235" containerID="867870b840279e987231e4c2b3fd57ae3a0d6af1a27a8b61f175ef51ea3434ab" exitCode=0 Nov 23 07:02:22 crc kubenswrapper[4906]: I1123 07:02:22.581173 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-k7tpf" event={"ID":"10146b05-1750-4ed1-ba76-ba7899743235","Type":"ContainerDied","Data":"867870b840279e987231e4c2b3fd57ae3a0d6af1a27a8b61f175ef51ea3434ab"} Nov 23 07:02:22 crc kubenswrapper[4906]: I1123 07:02:22.584779 4906 generic.go:334] "Generic (PLEG): container finished" podID="c45501a6-a479-43d6-ab52-805364aa1b21" containerID="9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1" exitCode=0 Nov 23 07:02:22 crc kubenswrapper[4906]: I1123 07:02:22.584847 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-975zp" event={"ID":"c45501a6-a479-43d6-ab52-805364aa1b21","Type":"ContainerDied","Data":"9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1"} Nov 23 07:02:23 crc kubenswrapper[4906]: I1123 07:02:23.597296 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-975zp" event={"ID":"c45501a6-a479-43d6-ab52-805364aa1b21","Type":"ContainerStarted","Data":"cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997"} Nov 23 07:02:23 crc kubenswrapper[4906]: I1123 07:02:23.627385 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-975zp" podStartSLOduration=3.218796663 podStartE2EDuration="5.627352415s" podCreationTimestamp="2025-11-23 07:02:18 +0000 UTC" firstStartedPulling="2025-11-23 07:02:20.565211897 +0000 UTC m=+756.078603210" lastFinishedPulling="2025-11-23 07:02:22.973767649 +0000 UTC m=+758.487158962" observedRunningTime="2025-11-23 07:02:23.622334044 +0000 UTC m=+759.135725387" watchObservedRunningTime="2025-11-23 07:02:23.627352415 +0000 UTC m=+759.140743758" Nov 23 07:02:23 crc kubenswrapper[4906]: I1123 07:02:23.947787 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.071174 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/10146b05-1750-4ed1-ba76-ba7899743235-crc-storage\") pod \"10146b05-1750-4ed1-ba76-ba7899743235\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.071270 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/10146b05-1750-4ed1-ba76-ba7899743235-node-mnt\") pod \"10146b05-1750-4ed1-ba76-ba7899743235\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.071404 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjtnw\" (UniqueName: \"kubernetes.io/projected/10146b05-1750-4ed1-ba76-ba7899743235-kube-api-access-xjtnw\") pod \"10146b05-1750-4ed1-ba76-ba7899743235\" (UID: \"10146b05-1750-4ed1-ba76-ba7899743235\") " Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.071845 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10146b05-1750-4ed1-ba76-ba7899743235-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "10146b05-1750-4ed1-ba76-ba7899743235" (UID: "10146b05-1750-4ed1-ba76-ba7899743235"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.081892 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10146b05-1750-4ed1-ba76-ba7899743235-kube-api-access-xjtnw" (OuterVolumeSpecName: "kube-api-access-xjtnw") pod "10146b05-1750-4ed1-ba76-ba7899743235" (UID: "10146b05-1750-4ed1-ba76-ba7899743235"). InnerVolumeSpecName "kube-api-access-xjtnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.098633 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10146b05-1750-4ed1-ba76-ba7899743235-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "10146b05-1750-4ed1-ba76-ba7899743235" (UID: "10146b05-1750-4ed1-ba76-ba7899743235"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.174204 4906 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/10146b05-1750-4ed1-ba76-ba7899743235-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.174273 4906 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/10146b05-1750-4ed1-ba76-ba7899743235-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.174302 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjtnw\" (UniqueName: \"kubernetes.io/projected/10146b05-1750-4ed1-ba76-ba7899743235-kube-api-access-xjtnw\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.606775 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k7tpf" Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.606757 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-k7tpf" event={"ID":"10146b05-1750-4ed1-ba76-ba7899743235","Type":"ContainerDied","Data":"db0555e00f4ae1894b519a0c7e60ebfaa5112a9c9472264ab4fe76826a01a50c"} Nov 23 07:02:24 crc kubenswrapper[4906]: I1123 07:02:24.606879 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db0555e00f4ae1894b519a0c7e60ebfaa5112a9c9472264ab4fe76826a01a50c" Nov 23 07:02:28 crc kubenswrapper[4906]: I1123 07:02:28.905842 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:28 crc kubenswrapper[4906]: I1123 07:02:28.908906 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:28 crc kubenswrapper[4906]: I1123 07:02:28.977581 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:29 crc kubenswrapper[4906]: I1123 07:02:29.719993 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:29 crc kubenswrapper[4906]: I1123 07:02:29.784272 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-975zp"] Nov 23 07:02:31 crc kubenswrapper[4906]: I1123 07:02:31.665050 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-975zp" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="registry-server" containerID="cri-o://cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997" gracePeriod=2 Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.127305 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.312317 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9qvj\" (UniqueName: \"kubernetes.io/projected/c45501a6-a479-43d6-ab52-805364aa1b21-kube-api-access-d9qvj\") pod \"c45501a6-a479-43d6-ab52-805364aa1b21\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.312507 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-utilities\") pod \"c45501a6-a479-43d6-ab52-805364aa1b21\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.312563 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-catalog-content\") pod \"c45501a6-a479-43d6-ab52-805364aa1b21\" (UID: \"c45501a6-a479-43d6-ab52-805364aa1b21\") " Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.313513 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-utilities" (OuterVolumeSpecName: "utilities") pod "c45501a6-a479-43d6-ab52-805364aa1b21" (UID: "c45501a6-a479-43d6-ab52-805364aa1b21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.320203 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c45501a6-a479-43d6-ab52-805364aa1b21-kube-api-access-d9qvj" (OuterVolumeSpecName: "kube-api-access-d9qvj") pod "c45501a6-a479-43d6-ab52-805364aa1b21" (UID: "c45501a6-a479-43d6-ab52-805364aa1b21"). InnerVolumeSpecName "kube-api-access-d9qvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.373648 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c45501a6-a479-43d6-ab52-805364aa1b21" (UID: "c45501a6-a479-43d6-ab52-805364aa1b21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.413885 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.413919 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45501a6-a479-43d6-ab52-805364aa1b21-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.413936 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9qvj\" (UniqueName: \"kubernetes.io/projected/c45501a6-a479-43d6-ab52-805364aa1b21-kube-api-access-d9qvj\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.676282 4906 generic.go:334] "Generic (PLEG): container finished" podID="c45501a6-a479-43d6-ab52-805364aa1b21" containerID="cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997" exitCode=0 Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.676356 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-975zp" event={"ID":"c45501a6-a479-43d6-ab52-805364aa1b21","Type":"ContainerDied","Data":"cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997"} Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.676384 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-975zp" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.676404 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-975zp" event={"ID":"c45501a6-a479-43d6-ab52-805364aa1b21","Type":"ContainerDied","Data":"5e9f813f2616677f0e726e7fa4ce720df7c0de2a95f2f1375189481ae6e6310a"} Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.676456 4906 scope.go:117] "RemoveContainer" containerID="cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.704819 4906 scope.go:117] "RemoveContainer" containerID="9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.723606 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-975zp"] Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.726340 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-975zp"] Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.741802 4906 scope.go:117] "RemoveContainer" containerID="65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.765524 4906 scope.go:117] "RemoveContainer" containerID="cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997" Nov 23 07:02:32 crc kubenswrapper[4906]: E1123 07:02:32.766198 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997\": container with ID starting with cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997 not found: ID does not exist" containerID="cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.766255 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997"} err="failed to get container status \"cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997\": rpc error: code = NotFound desc = could not find container \"cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997\": container with ID starting with cac1c3232c7e4e4950c8a9a9badb9f310839015baef4d9c899f084e39e59d997 not found: ID does not exist" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.766287 4906 scope.go:117] "RemoveContainer" containerID="9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1" Nov 23 07:02:32 crc kubenswrapper[4906]: E1123 07:02:32.766632 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1\": container with ID starting with 9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1 not found: ID does not exist" containerID="9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.766671 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1"} err="failed to get container status \"9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1\": rpc error: code = NotFound desc = could not find container \"9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1\": container with ID starting with 9a131ba0d5933a162c05034c3a23230de13273cf5e691c22f2ab9732c7915db1 not found: ID does not exist" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.766726 4906 scope.go:117] "RemoveContainer" containerID="65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9" Nov 23 07:02:32 crc kubenswrapper[4906]: E1123 07:02:32.767145 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9\": container with ID starting with 65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9 not found: ID does not exist" containerID="65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.767177 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9"} err="failed to get container status \"65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9\": rpc error: code = NotFound desc = could not find container \"65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9\": container with ID starting with 65a261382c553f2a8058ce1a113147fe49e7eec57f960331325c8ac3abc2dde9 not found: ID does not exist" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.899636 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg"] Nov 23 07:02:32 crc kubenswrapper[4906]: E1123 07:02:32.899994 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="extract-content" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.900014 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="extract-content" Nov 23 07:02:32 crc kubenswrapper[4906]: E1123 07:02:32.900024 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="registry-server" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.900031 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="registry-server" Nov 23 07:02:32 crc kubenswrapper[4906]: E1123 07:02:32.900042 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10146b05-1750-4ed1-ba76-ba7899743235" containerName="storage" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.900049 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="10146b05-1750-4ed1-ba76-ba7899743235" containerName="storage" Nov 23 07:02:32 crc kubenswrapper[4906]: E1123 07:02:32.900059 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="extract-utilities" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.900067 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="extract-utilities" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.900157 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" containerName="registry-server" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.900170 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="10146b05-1750-4ed1-ba76-ba7899743235" containerName="storage" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.900995 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.904710 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 23 07:02:32 crc kubenswrapper[4906]: I1123 07:02:32.916630 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg"] Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.024421 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.024588 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.024659 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk5kl\" (UniqueName: \"kubernetes.io/projected/86ca56c2-497b-498d-b69f-5f80f17c1765-kube-api-access-vk5kl\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.126012 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.126106 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk5kl\" (UniqueName: \"kubernetes.io/projected/86ca56c2-497b-498d-b69f-5f80f17c1765-kube-api-access-vk5kl\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.126151 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.126822 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.126995 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.151593 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk5kl\" (UniqueName: \"kubernetes.io/projected/86ca56c2-497b-498d-b69f-5f80f17c1765-kube-api-access-vk5kl\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.216372 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.374838 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c45501a6-a479-43d6-ab52-805364aa1b21" path="/var/lib/kubelet/pods/c45501a6-a479-43d6-ab52-805364aa1b21/volumes" Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.466333 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg"] Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.685914 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" event={"ID":"86ca56c2-497b-498d-b69f-5f80f17c1765","Type":"ContainerStarted","Data":"b20630d43846a46349a76daacd8d32155b36a1a0c06a43f8e0f719f81e4ede50"} Nov 23 07:02:33 crc kubenswrapper[4906]: I1123 07:02:33.685958 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" event={"ID":"86ca56c2-497b-498d-b69f-5f80f17c1765","Type":"ContainerStarted","Data":"cbae8becbb148c5c3f4c0eb8ab476bd968c59c9ff9d82eddcae7505dd4a5bede"} Nov 23 07:02:34 crc kubenswrapper[4906]: I1123 07:02:34.701829 4906 generic.go:334] "Generic (PLEG): container finished" podID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerID="b20630d43846a46349a76daacd8d32155b36a1a0c06a43f8e0f719f81e4ede50" exitCode=0 Nov 23 07:02:34 crc kubenswrapper[4906]: I1123 07:02:34.701962 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" event={"ID":"86ca56c2-497b-498d-b69f-5f80f17c1765","Type":"ContainerDied","Data":"b20630d43846a46349a76daacd8d32155b36a1a0c06a43f8e0f719f81e4ede50"} Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.453462 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cl9kp"] Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.459851 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.473425 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cl9kp"] Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.595868 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-catalog-content\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.595989 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb7px\" (UniqueName: \"kubernetes.io/projected/21ceb6c0-5af1-4509-8538-0335cc204442-kube-api-access-rb7px\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.596773 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-utilities\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.698341 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-utilities\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.698417 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-catalog-content\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.698451 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb7px\" (UniqueName: \"kubernetes.io/projected/21ceb6c0-5af1-4509-8538-0335cc204442-kube-api-access-rb7px\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.699312 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-utilities\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.699373 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-catalog-content\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.719672 4906 generic.go:334] "Generic (PLEG): container finished" podID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerID="1daede85bd648d8ac7ac5d96d1a1ccb1b0c9433aced9ef79b20ab08f1d74276b" exitCode=0 Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.719742 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" event={"ID":"86ca56c2-497b-498d-b69f-5f80f17c1765","Type":"ContainerDied","Data":"1daede85bd648d8ac7ac5d96d1a1ccb1b0c9433aced9ef79b20ab08f1d74276b"} Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.733729 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb7px\" (UniqueName: \"kubernetes.io/projected/21ceb6c0-5af1-4509-8538-0335cc204442-kube-api-access-rb7px\") pod \"redhat-operators-cl9kp\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.784496 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:36 crc kubenswrapper[4906]: I1123 07:02:36.992002 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cl9kp"] Nov 23 07:02:37 crc kubenswrapper[4906]: W1123 07:02:36.999932 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21ceb6c0_5af1_4509_8538_0335cc204442.slice/crio-c608400e2eb3b4c586478f931c7dd123ff43a116a184627d4eef8fb3ac6452d0 WatchSource:0}: Error finding container c608400e2eb3b4c586478f931c7dd123ff43a116a184627d4eef8fb3ac6452d0: Status 404 returned error can't find the container with id c608400e2eb3b4c586478f931c7dd123ff43a116a184627d4eef8fb3ac6452d0 Nov 23 07:02:37 crc kubenswrapper[4906]: I1123 07:02:37.730653 4906 generic.go:334] "Generic (PLEG): container finished" podID="21ceb6c0-5af1-4509-8538-0335cc204442" containerID="9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041" exitCode=0 Nov 23 07:02:37 crc kubenswrapper[4906]: I1123 07:02:37.730738 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl9kp" event={"ID":"21ceb6c0-5af1-4509-8538-0335cc204442","Type":"ContainerDied","Data":"9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041"} Nov 23 07:02:37 crc kubenswrapper[4906]: I1123 07:02:37.730817 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl9kp" event={"ID":"21ceb6c0-5af1-4509-8538-0335cc204442","Type":"ContainerStarted","Data":"c608400e2eb3b4c586478f931c7dd123ff43a116a184627d4eef8fb3ac6452d0"} Nov 23 07:02:37 crc kubenswrapper[4906]: I1123 07:02:37.733476 4906 generic.go:334] "Generic (PLEG): container finished" podID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerID="51a175facf504041d701a15bcdd82537a3351612e0591a66aaf949b99c74e38a" exitCode=0 Nov 23 07:02:37 crc kubenswrapper[4906]: I1123 07:02:37.733552 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" event={"ID":"86ca56c2-497b-498d-b69f-5f80f17c1765","Type":"ContainerDied","Data":"51a175facf504041d701a15bcdd82537a3351612e0591a66aaf949b99c74e38a"} Nov 23 07:02:38 crc kubenswrapper[4906]: I1123 07:02:38.483077 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mpt2b" Nov 23 07:02:38 crc kubenswrapper[4906]: I1123 07:02:38.743491 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl9kp" event={"ID":"21ceb6c0-5af1-4509-8538-0335cc204442","Type":"ContainerStarted","Data":"618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680"} Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.061624 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.139030 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-bundle\") pod \"86ca56c2-497b-498d-b69f-5f80f17c1765\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.139659 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-bundle" (OuterVolumeSpecName: "bundle") pod "86ca56c2-497b-498d-b69f-5f80f17c1765" (UID: "86ca56c2-497b-498d-b69f-5f80f17c1765"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.240616 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk5kl\" (UniqueName: \"kubernetes.io/projected/86ca56c2-497b-498d-b69f-5f80f17c1765-kube-api-access-vk5kl\") pod \"86ca56c2-497b-498d-b69f-5f80f17c1765\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.240755 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-util\") pod \"86ca56c2-497b-498d-b69f-5f80f17c1765\" (UID: \"86ca56c2-497b-498d-b69f-5f80f17c1765\") " Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.242102 4906 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.272103 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ca56c2-497b-498d-b69f-5f80f17c1765-kube-api-access-vk5kl" (OuterVolumeSpecName: "kube-api-access-vk5kl") pod "86ca56c2-497b-498d-b69f-5f80f17c1765" (UID: "86ca56c2-497b-498d-b69f-5f80f17c1765"). InnerVolumeSpecName "kube-api-access-vk5kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.343872 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk5kl\" (UniqueName: \"kubernetes.io/projected/86ca56c2-497b-498d-b69f-5f80f17c1765-kube-api-access-vk5kl\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.470448 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-util" (OuterVolumeSpecName: "util") pod "86ca56c2-497b-498d-b69f-5f80f17c1765" (UID: "86ca56c2-497b-498d-b69f-5f80f17c1765"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.547724 4906 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86ca56c2-497b-498d-b69f-5f80f17c1765-util\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.757010 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.757465 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg" event={"ID":"86ca56c2-497b-498d-b69f-5f80f17c1765","Type":"ContainerDied","Data":"cbae8becbb148c5c3f4c0eb8ab476bd968c59c9ff9d82eddcae7505dd4a5bede"} Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.757527 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbae8becbb148c5c3f4c0eb8ab476bd968c59c9ff9d82eddcae7505dd4a5bede" Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.761426 4906 generic.go:334] "Generic (PLEG): container finished" podID="21ceb6c0-5af1-4509-8538-0335cc204442" containerID="618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680" exitCode=0 Nov 23 07:02:39 crc kubenswrapper[4906]: I1123 07:02:39.761475 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl9kp" event={"ID":"21ceb6c0-5af1-4509-8538-0335cc204442","Type":"ContainerDied","Data":"618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680"} Nov 23 07:02:40 crc kubenswrapper[4906]: I1123 07:02:40.768391 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl9kp" event={"ID":"21ceb6c0-5af1-4509-8538-0335cc204442","Type":"ContainerStarted","Data":"50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38"} Nov 23 07:02:40 crc kubenswrapper[4906]: I1123 07:02:40.786935 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cl9kp" podStartSLOduration=2.330983576 podStartE2EDuration="4.786905846s" podCreationTimestamp="2025-11-23 07:02:36 +0000 UTC" firstStartedPulling="2025-11-23 07:02:37.733203189 +0000 UTC m=+773.246594522" lastFinishedPulling="2025-11-23 07:02:40.189125479 +0000 UTC m=+775.702516792" observedRunningTime="2025-11-23 07:02:40.784720248 +0000 UTC m=+776.298111551" watchObservedRunningTime="2025-11-23 07:02:40.786905846 +0000 UTC m=+776.300297159" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.170062 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-9qw8j"] Nov 23 07:02:41 crc kubenswrapper[4906]: E1123 07:02:41.170347 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerName="extract" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.170362 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerName="extract" Nov 23 07:02:41 crc kubenswrapper[4906]: E1123 07:02:41.170374 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerName="util" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.170394 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerName="util" Nov 23 07:02:41 crc kubenswrapper[4906]: E1123 07:02:41.170425 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerName="pull" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.170433 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerName="pull" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.170564 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ca56c2-497b-498d-b69f-5f80f17c1765" containerName="extract" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.171090 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.173819 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.174079 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vsrf2" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.174210 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.186918 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-9qw8j"] Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.274017 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhvc6\" (UniqueName: \"kubernetes.io/projected/30760439-eebd-40db-a8fc-7c4c355b6988-kube-api-access-rhvc6\") pod \"nmstate-operator-557fdffb88-9qw8j\" (UID: \"30760439-eebd-40db-a8fc-7c4c355b6988\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.376013 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhvc6\" (UniqueName: \"kubernetes.io/projected/30760439-eebd-40db-a8fc-7c4c355b6988-kube-api-access-rhvc6\") pod \"nmstate-operator-557fdffb88-9qw8j\" (UID: \"30760439-eebd-40db-a8fc-7c4c355b6988\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.405968 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhvc6\" (UniqueName: \"kubernetes.io/projected/30760439-eebd-40db-a8fc-7c4c355b6988-kube-api-access-rhvc6\") pod \"nmstate-operator-557fdffb88-9qw8j\" (UID: \"30760439-eebd-40db-a8fc-7c4c355b6988\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.488387 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" Nov 23 07:02:41 crc kubenswrapper[4906]: I1123 07:02:41.816040 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-9qw8j"] Nov 23 07:02:41 crc kubenswrapper[4906]: W1123 07:02:41.839153 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30760439_eebd_40db_a8fc_7c4c355b6988.slice/crio-1717fe24dc816da163a6bbabb32d4ba1722981de425e4e2d39199474f74e483e WatchSource:0}: Error finding container 1717fe24dc816da163a6bbabb32d4ba1722981de425e4e2d39199474f74e483e: Status 404 returned error can't find the container with id 1717fe24dc816da163a6bbabb32d4ba1722981de425e4e2d39199474f74e483e Nov 23 07:02:42 crc kubenswrapper[4906]: I1123 07:02:42.794838 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" event={"ID":"30760439-eebd-40db-a8fc-7c4c355b6988","Type":"ContainerStarted","Data":"1717fe24dc816da163a6bbabb32d4ba1722981de425e4e2d39199474f74e483e"} Nov 23 07:02:44 crc kubenswrapper[4906]: I1123 07:02:44.810564 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" event={"ID":"30760439-eebd-40db-a8fc-7c4c355b6988","Type":"ContainerStarted","Data":"844affb3ef09b4ea18f6fd7939eedfb0d35674d49a7ed4526986baa56119f219"} Nov 23 07:02:44 crc kubenswrapper[4906]: I1123 07:02:44.829267 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-9qw8j" podStartSLOduration=1.629187445 podStartE2EDuration="3.829247173s" podCreationTimestamp="2025-11-23 07:02:41 +0000 UTC" firstStartedPulling="2025-11-23 07:02:41.841728377 +0000 UTC m=+777.355119680" lastFinishedPulling="2025-11-23 07:02:44.041788105 +0000 UTC m=+779.555179408" observedRunningTime="2025-11-23 07:02:44.825443753 +0000 UTC m=+780.338835066" watchObservedRunningTime="2025-11-23 07:02:44.829247173 +0000 UTC m=+780.342638476" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.659644 4906 scope.go:117] "RemoveContainer" containerID="9fb1d0991fa29985db49339d850ab1b8340a5ee32b9ec2b6a4145e5b22423b28" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.820848 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lm9xt_4f75f165-ef56-40be-bfd4-1843de92b356/kube-multus/2.log" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.874916 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n"] Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.876668 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.879057 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-b5hgv" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.892288 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc"] Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.893145 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.898517 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.902306 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-87hwm"] Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.903664 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.936443 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n"] Nov 23 07:02:45 crc kubenswrapper[4906]: I1123 07:02:45.942672 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc"] Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.051648 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq"] Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.055359 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.059877 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060245 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9ff9\" (UniqueName: \"kubernetes.io/projected/8fff20a6-d9f6-451e-9291-56b3d7c27a4d-kube-api-access-g9ff9\") pod \"nmstate-webhook-6b89b748d8-tglkc\" (UID: \"8fff20a6-d9f6-451e-9291-56b3d7c27a4d\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060316 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-nmstate-lock\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060359 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-ovs-socket\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060388 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-dbus-socket\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060471 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8fff20a6-d9f6-451e-9291-56b3d7c27a4d-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-tglkc\" (UID: \"8fff20a6-d9f6-451e-9291-56b3d7c27a4d\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060512 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpm5l\" (UniqueName: \"kubernetes.io/projected/b16b1f54-fe15-490f-9f7b-d77ca6708c2e-kube-api-access-wpm5l\") pod \"nmstate-metrics-5dcf9c57c5-hww9n\" (UID: \"b16b1f54-fe15-490f-9f7b-d77ca6708c2e\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060596 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060619 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds46n\" (UniqueName: \"kubernetes.io/projected/a702c630-d383-4ed0-82d5-07e8e074443a-kube-api-access-ds46n\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.060658 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-hmk5x" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.061030 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq"] Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.162010 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb2jf\" (UniqueName: \"kubernetes.io/projected/acf85f16-cd96-4cc5-b271-16444206d205-kube-api-access-hb2jf\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.162066 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8fff20a6-d9f6-451e-9291-56b3d7c27a4d-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-tglkc\" (UID: \"8fff20a6-d9f6-451e-9291-56b3d7c27a4d\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.162095 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpm5l\" (UniqueName: \"kubernetes.io/projected/b16b1f54-fe15-490f-9f7b-d77ca6708c2e-kube-api-access-wpm5l\") pod \"nmstate-metrics-5dcf9c57c5-hww9n\" (UID: \"b16b1f54-fe15-490f-9f7b-d77ca6708c2e\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163216 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds46n\" (UniqueName: \"kubernetes.io/projected/a702c630-d383-4ed0-82d5-07e8e074443a-kube-api-access-ds46n\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163273 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9ff9\" (UniqueName: \"kubernetes.io/projected/8fff20a6-d9f6-451e-9291-56b3d7c27a4d-kube-api-access-g9ff9\") pod \"nmstate-webhook-6b89b748d8-tglkc\" (UID: \"8fff20a6-d9f6-451e-9291-56b3d7c27a4d\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163323 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-nmstate-lock\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163386 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-ovs-socket\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163413 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/acf85f16-cd96-4cc5-b271-16444206d205-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163444 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-dbus-socket\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163473 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/acf85f16-cd96-4cc5-b271-16444206d205-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163487 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-nmstate-lock\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163572 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-ovs-socket\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.163923 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a702c630-d383-4ed0-82d5-07e8e074443a-dbus-socket\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.172556 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8fff20a6-d9f6-451e-9291-56b3d7c27a4d-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-tglkc\" (UID: \"8fff20a6-d9f6-451e-9291-56b3d7c27a4d\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.190152 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpm5l\" (UniqueName: \"kubernetes.io/projected/b16b1f54-fe15-490f-9f7b-d77ca6708c2e-kube-api-access-wpm5l\") pod \"nmstate-metrics-5dcf9c57c5-hww9n\" (UID: \"b16b1f54-fe15-490f-9f7b-d77ca6708c2e\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.192255 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.195778 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9ff9\" (UniqueName: \"kubernetes.io/projected/8fff20a6-d9f6-451e-9291-56b3d7c27a4d-kube-api-access-g9ff9\") pod \"nmstate-webhook-6b89b748d8-tglkc\" (UID: \"8fff20a6-d9f6-451e-9291-56b3d7c27a4d\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.199058 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds46n\" (UniqueName: \"kubernetes.io/projected/a702c630-d383-4ed0-82d5-07e8e074443a-kube-api-access-ds46n\") pod \"nmstate-handler-87hwm\" (UID: \"a702c630-d383-4ed0-82d5-07e8e074443a\") " pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.207966 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.221059 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.258825 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6d7bcd6f86-jhh4p"] Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.259970 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.266033 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/acf85f16-cd96-4cc5-b271-16444206d205-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.266090 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb2jf\" (UniqueName: \"kubernetes.io/projected/acf85f16-cd96-4cc5-b271-16444206d205-kube-api-access-hb2jf\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.266152 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/acf85f16-cd96-4cc5-b271-16444206d205-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: E1123 07:02:46.266277 4906 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 23 07:02:46 crc kubenswrapper[4906]: E1123 07:02:46.266337 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/acf85f16-cd96-4cc5-b271-16444206d205-plugin-serving-cert podName:acf85f16-cd96-4cc5-b271-16444206d205 nodeName:}" failed. No retries permitted until 2025-11-23 07:02:46.766318993 +0000 UTC m=+782.279710296 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/acf85f16-cd96-4cc5-b271-16444206d205-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-qgsjq" (UID: "acf85f16-cd96-4cc5-b271-16444206d205") : secret "plugin-serving-cert" not found Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.275636 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/acf85f16-cd96-4cc5-b271-16444206d205-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.278296 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d7bcd6f86-jhh4p"] Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.310927 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb2jf\" (UniqueName: \"kubernetes.io/projected/acf85f16-cd96-4cc5-b271-16444206d205-kube-api-access-hb2jf\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.366942 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-oauth-serving-cert\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.367022 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-trusted-ca-bundle\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.367277 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-service-ca\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.367301 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f373ed95-6499-4faa-83d3-e5d716e045ed-console-oauth-config\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.367320 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkbtm\" (UniqueName: \"kubernetes.io/projected/f373ed95-6499-4faa-83d3-e5d716e045ed-kube-api-access-pkbtm\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.367338 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f373ed95-6499-4faa-83d3-e5d716e045ed-console-serving-cert\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.367412 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-console-config\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.468673 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-oauth-serving-cert\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469123 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-trusted-ca-bundle\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469169 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-service-ca\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469187 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f373ed95-6499-4faa-83d3-e5d716e045ed-console-oauth-config\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469208 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkbtm\" (UniqueName: \"kubernetes.io/projected/f373ed95-6499-4faa-83d3-e5d716e045ed-kube-api-access-pkbtm\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469226 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f373ed95-6499-4faa-83d3-e5d716e045ed-console-serving-cert\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469266 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-console-config\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469870 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-oauth-serving-cert\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.469907 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-console-config\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.471848 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-trusted-ca-bundle\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.474052 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f373ed95-6499-4faa-83d3-e5d716e045ed-service-ca\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.476448 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f373ed95-6499-4faa-83d3-e5d716e045ed-console-serving-cert\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.476719 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f373ed95-6499-4faa-83d3-e5d716e045ed-console-oauth-config\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.490504 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkbtm\" (UniqueName: \"kubernetes.io/projected/f373ed95-6499-4faa-83d3-e5d716e045ed-kube-api-access-pkbtm\") pod \"console-6d7bcd6f86-jhh4p\" (UID: \"f373ed95-6499-4faa-83d3-e5d716e045ed\") " pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.560994 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc"] Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.591854 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.669003 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n"] Nov 23 07:02:46 crc kubenswrapper[4906]: W1123 07:02:46.678963 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb16b1f54_fe15_490f_9f7b_d77ca6708c2e.slice/crio-c46c0ae80f59bacf1f71fd589e79bf5090cce6aef63ae4fca50c0778d41da440 WatchSource:0}: Error finding container c46c0ae80f59bacf1f71fd589e79bf5090cce6aef63ae4fca50c0778d41da440: Status 404 returned error can't find the container with id c46c0ae80f59bacf1f71fd589e79bf5090cce6aef63ae4fca50c0778d41da440 Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.773123 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/acf85f16-cd96-4cc5-b271-16444206d205-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.778698 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/acf85f16-cd96-4cc5-b271-16444206d205-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-qgsjq\" (UID: \"acf85f16-cd96-4cc5-b271-16444206d205\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.785979 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.786403 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.835833 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-87hwm" event={"ID":"a702c630-d383-4ed0-82d5-07e8e074443a","Type":"ContainerStarted","Data":"b97de7dc1bd3bff651df22858c99837d261e8c8f873d66d06f929eb99704db66"} Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.836324 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d7bcd6f86-jhh4p"] Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.850000 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" event={"ID":"b16b1f54-fe15-490f-9f7b-d77ca6708c2e","Type":"ContainerStarted","Data":"c46c0ae80f59bacf1f71fd589e79bf5090cce6aef63ae4fca50c0778d41da440"} Nov 23 07:02:46 crc kubenswrapper[4906]: W1123 07:02:46.852220 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf373ed95_6499_4faa_83d3_e5d716e045ed.slice/crio-ff37ee7bbdb2a85506c31b4d74a43ff69b0fae7f4a540db6d2ed2c993f684a7c WatchSource:0}: Error finding container ff37ee7bbdb2a85506c31b4d74a43ff69b0fae7f4a540db6d2ed2c993f684a7c: Status 404 returned error can't find the container with id ff37ee7bbdb2a85506c31b4d74a43ff69b0fae7f4a540db6d2ed2c993f684a7c Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.852408 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" event={"ID":"8fff20a6-d9f6-451e-9291-56b3d7c27a4d","Type":"ContainerStarted","Data":"5ae80c5810682b899dd1cfe1ff587ebda99e4d5521bcf4a9c7a24ab5cfde5059"} Nov 23 07:02:46 crc kubenswrapper[4906]: I1123 07:02:46.976120 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" Nov 23 07:02:47 crc kubenswrapper[4906]: I1123 07:02:47.193377 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq"] Nov 23 07:02:47 crc kubenswrapper[4906]: I1123 07:02:47.838359 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cl9kp" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="registry-server" probeResult="failure" output=< Nov 23 07:02:47 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 07:02:47 crc kubenswrapper[4906]: > Nov 23 07:02:47 crc kubenswrapper[4906]: I1123 07:02:47.865543 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d7bcd6f86-jhh4p" event={"ID":"f373ed95-6499-4faa-83d3-e5d716e045ed","Type":"ContainerStarted","Data":"b0460093a181149b129c6ddd7c78ce5412f053a1f0cb01fd855e399790e7eb14"} Nov 23 07:02:47 crc kubenswrapper[4906]: I1123 07:02:47.865637 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d7bcd6f86-jhh4p" event={"ID":"f373ed95-6499-4faa-83d3-e5d716e045ed","Type":"ContainerStarted","Data":"ff37ee7bbdb2a85506c31b4d74a43ff69b0fae7f4a540db6d2ed2c993f684a7c"} Nov 23 07:02:47 crc kubenswrapper[4906]: I1123 07:02:47.868519 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" event={"ID":"acf85f16-cd96-4cc5-b271-16444206d205","Type":"ContainerStarted","Data":"6122b90fd6ae432213a5a949b2c799b725989dec7e24a5318206985abcf484dc"} Nov 23 07:02:47 crc kubenswrapper[4906]: I1123 07:02:47.889067 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6d7bcd6f86-jhh4p" podStartSLOduration=1.889041449 podStartE2EDuration="1.889041449s" podCreationTimestamp="2025-11-23 07:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:02:47.882847708 +0000 UTC m=+783.396239001" watchObservedRunningTime="2025-11-23 07:02:47.889041449 +0000 UTC m=+783.402432752" Nov 23 07:02:49 crc kubenswrapper[4906]: I1123 07:02:49.894500 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-87hwm" event={"ID":"a702c630-d383-4ed0-82d5-07e8e074443a","Type":"ContainerStarted","Data":"16403ec0ad930a170f3f473afa43a01127f55cdd3a51c1a42f5116aea90c52ef"} Nov 23 07:02:49 crc kubenswrapper[4906]: I1123 07:02:49.895897 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:49 crc kubenswrapper[4906]: I1123 07:02:49.898654 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" event={"ID":"b16b1f54-fe15-490f-9f7b-d77ca6708c2e","Type":"ContainerStarted","Data":"11caeec67aa57110315ec66fcdf0d50c02ce4c490f00f9e788b05787e9efc7bb"} Nov 23 07:02:49 crc kubenswrapper[4906]: I1123 07:02:49.900514 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" event={"ID":"8fff20a6-d9f6-451e-9291-56b3d7c27a4d","Type":"ContainerStarted","Data":"4fe50abf28b34e59f76ed2649c2eff95a0902c52063fb09414b9208637b7457d"} Nov 23 07:02:49 crc kubenswrapper[4906]: I1123 07:02:49.901031 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:02:49 crc kubenswrapper[4906]: I1123 07:02:49.910747 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-87hwm" podStartSLOduration=2.333913418 podStartE2EDuration="4.910725432s" podCreationTimestamp="2025-11-23 07:02:45 +0000 UTC" firstStartedPulling="2025-11-23 07:02:46.298620478 +0000 UTC m=+781.812011781" lastFinishedPulling="2025-11-23 07:02:48.875432482 +0000 UTC m=+784.388823795" observedRunningTime="2025-11-23 07:02:49.908056432 +0000 UTC m=+785.421447765" watchObservedRunningTime="2025-11-23 07:02:49.910725432 +0000 UTC m=+785.424116735" Nov 23 07:02:49 crc kubenswrapper[4906]: I1123 07:02:49.927824 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" podStartSLOduration=2.2575505209999998 podStartE2EDuration="4.927804658s" podCreationTimestamp="2025-11-23 07:02:45 +0000 UTC" firstStartedPulling="2025-11-23 07:02:46.567544422 +0000 UTC m=+782.080935725" lastFinishedPulling="2025-11-23 07:02:49.237798529 +0000 UTC m=+784.751189862" observedRunningTime="2025-11-23 07:02:49.925738375 +0000 UTC m=+785.439129698" watchObservedRunningTime="2025-11-23 07:02:49.927804658 +0000 UTC m=+785.441195961" Nov 23 07:02:50 crc kubenswrapper[4906]: I1123 07:02:50.946099 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:02:50 crc kubenswrapper[4906]: I1123 07:02:50.946472 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:02:51 crc kubenswrapper[4906]: I1123 07:02:51.919482 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" event={"ID":"acf85f16-cd96-4cc5-b271-16444206d205","Type":"ContainerStarted","Data":"8f9717898d81604c48f4bef069bfd2dd753eaf14f7e822dfd5504ce2fd8a28c6"} Nov 23 07:02:53 crc kubenswrapper[4906]: I1123 07:02:53.939195 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" event={"ID":"b16b1f54-fe15-490f-9f7b-d77ca6708c2e","Type":"ContainerStarted","Data":"b048a505779a1e3b853f79cd622ce22d77aaed682c486783be0ecf0ca0883dda"} Nov 23 07:02:53 crc kubenswrapper[4906]: I1123 07:02:53.959226 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-hww9n" podStartSLOduration=2.428930015 podStartE2EDuration="8.95919841s" podCreationTimestamp="2025-11-23 07:02:45 +0000 UTC" firstStartedPulling="2025-11-23 07:02:46.681786071 +0000 UTC m=+782.195177374" lastFinishedPulling="2025-11-23 07:02:53.212054466 +0000 UTC m=+788.725445769" observedRunningTime="2025-11-23 07:02:53.956253012 +0000 UTC m=+789.469644345" watchObservedRunningTime="2025-11-23 07:02:53.95919841 +0000 UTC m=+789.472589713" Nov 23 07:02:53 crc kubenswrapper[4906]: I1123 07:02:53.960244 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-qgsjq" podStartSLOduration=4.867027747 podStartE2EDuration="7.960234267s" podCreationTimestamp="2025-11-23 07:02:46 +0000 UTC" firstStartedPulling="2025-11-23 07:02:47.200020087 +0000 UTC m=+782.713411390" lastFinishedPulling="2025-11-23 07:02:50.293226567 +0000 UTC m=+785.806617910" observedRunningTime="2025-11-23 07:02:52.949902769 +0000 UTC m=+788.463294072" watchObservedRunningTime="2025-11-23 07:02:53.960234267 +0000 UTC m=+789.473625570" Nov 23 07:02:56 crc kubenswrapper[4906]: I1123 07:02:56.261776 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-87hwm" Nov 23 07:02:56 crc kubenswrapper[4906]: I1123 07:02:56.592122 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:56 crc kubenswrapper[4906]: I1123 07:02:56.592612 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:56 crc kubenswrapper[4906]: I1123 07:02:56.598515 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:56 crc kubenswrapper[4906]: I1123 07:02:56.850054 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:56 crc kubenswrapper[4906]: I1123 07:02:56.901097 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:56 crc kubenswrapper[4906]: I1123 07:02:56.968328 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6d7bcd6f86-jhh4p" Nov 23 07:02:57 crc kubenswrapper[4906]: I1123 07:02:57.032342 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wp8fn"] Nov 23 07:02:57 crc kubenswrapper[4906]: I1123 07:02:57.103550 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cl9kp"] Nov 23 07:02:57 crc kubenswrapper[4906]: I1123 07:02:57.969995 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cl9kp" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="registry-server" containerID="cri-o://50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38" gracePeriod=2 Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.829521 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.977591 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-catalog-content\") pod \"21ceb6c0-5af1-4509-8538-0335cc204442\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.977707 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb7px\" (UniqueName: \"kubernetes.io/projected/21ceb6c0-5af1-4509-8538-0335cc204442-kube-api-access-rb7px\") pod \"21ceb6c0-5af1-4509-8538-0335cc204442\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.977878 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-utilities\") pod \"21ceb6c0-5af1-4509-8538-0335cc204442\" (UID: \"21ceb6c0-5af1-4509-8538-0335cc204442\") " Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.979010 4906 generic.go:334] "Generic (PLEG): container finished" podID="21ceb6c0-5af1-4509-8538-0335cc204442" containerID="50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38" exitCode=0 Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.979096 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl9kp" event={"ID":"21ceb6c0-5af1-4509-8538-0335cc204442","Type":"ContainerDied","Data":"50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38"} Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.979136 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl9kp" event={"ID":"21ceb6c0-5af1-4509-8538-0335cc204442","Type":"ContainerDied","Data":"c608400e2eb3b4c586478f931c7dd123ff43a116a184627d4eef8fb3ac6452d0"} Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.979158 4906 scope.go:117] "RemoveContainer" containerID="50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38" Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.979303 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl9kp" Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.979906 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-utilities" (OuterVolumeSpecName: "utilities") pod "21ceb6c0-5af1-4509-8538-0335cc204442" (UID: "21ceb6c0-5af1-4509-8538-0335cc204442"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:58 crc kubenswrapper[4906]: I1123 07:02:58.985102 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21ceb6c0-5af1-4509-8538-0335cc204442-kube-api-access-rb7px" (OuterVolumeSpecName: "kube-api-access-rb7px") pod "21ceb6c0-5af1-4509-8538-0335cc204442" (UID: "21ceb6c0-5af1-4509-8538-0335cc204442"). InnerVolumeSpecName "kube-api-access-rb7px". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.028625 4906 scope.go:117] "RemoveContainer" containerID="618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.055258 4906 scope.go:117] "RemoveContainer" containerID="9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.080188 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb7px\" (UniqueName: \"kubernetes.io/projected/21ceb6c0-5af1-4509-8538-0335cc204442-kube-api-access-rb7px\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.080274 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.088401 4906 scope.go:117] "RemoveContainer" containerID="50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38" Nov 23 07:02:59 crc kubenswrapper[4906]: E1123 07:02:59.089349 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38\": container with ID starting with 50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38 not found: ID does not exist" containerID="50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.089430 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38"} err="failed to get container status \"50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38\": rpc error: code = NotFound desc = could not find container \"50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38\": container with ID starting with 50642cae09b36db538848d321a5f4bd775a7c2ec3cf12dbefdee08bc3c29da38 not found: ID does not exist" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.089486 4906 scope.go:117] "RemoveContainer" containerID="618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680" Nov 23 07:02:59 crc kubenswrapper[4906]: E1123 07:02:59.090487 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680\": container with ID starting with 618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680 not found: ID does not exist" containerID="618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.090564 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680"} err="failed to get container status \"618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680\": rpc error: code = NotFound desc = could not find container \"618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680\": container with ID starting with 618755bf1fc20e33f978cdb9d1c43467118018dbdd0e012e3fdb32bf31117680 not found: ID does not exist" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.090620 4906 scope.go:117] "RemoveContainer" containerID="9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041" Nov 23 07:02:59 crc kubenswrapper[4906]: E1123 07:02:59.091161 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041\": container with ID starting with 9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041 not found: ID does not exist" containerID="9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.091221 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041"} err="failed to get container status \"9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041\": rpc error: code = NotFound desc = could not find container \"9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041\": container with ID starting with 9e277e4df04edbe467d30723d768663f6361817f69a033addc43b3e51a30d041 not found: ID does not exist" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.099062 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21ceb6c0-5af1-4509-8538-0335cc204442" (UID: "21ceb6c0-5af1-4509-8538-0335cc204442"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.182462 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21ceb6c0-5af1-4509-8538-0335cc204442-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.333130 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cl9kp"] Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.336675 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cl9kp"] Nov 23 07:02:59 crc kubenswrapper[4906]: I1123 07:02:59.382973 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" path="/var/lib/kubelet/pods/21ceb6c0-5af1-4509-8538-0335cc204442/volumes" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.491266 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wwp7w"] Nov 23 07:03:03 crc kubenswrapper[4906]: E1123 07:03:03.493465 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="extract-content" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.493537 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="extract-content" Nov 23 07:03:03 crc kubenswrapper[4906]: E1123 07:03:03.493576 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="extract-utilities" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.493596 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="extract-utilities" Nov 23 07:03:03 crc kubenswrapper[4906]: E1123 07:03:03.493636 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="registry-server" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.493654 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="registry-server" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.493989 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="21ceb6c0-5af1-4509-8538-0335cc204442" containerName="registry-server" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.496015 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.509315 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwp7w"] Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.674914 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-utilities\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.675060 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkh7h\" (UniqueName: \"kubernetes.io/projected/f47ba087-daee-498c-9e84-1a3e58ef6d3e-kube-api-access-jkh7h\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.675131 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-catalog-content\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.776723 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-utilities\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.776783 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkh7h\" (UniqueName: \"kubernetes.io/projected/f47ba087-daee-498c-9e84-1a3e58ef6d3e-kube-api-access-jkh7h\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.776817 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-catalog-content\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.777362 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-catalog-content\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.777670 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-utilities\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.802529 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkh7h\" (UniqueName: \"kubernetes.io/projected/f47ba087-daee-498c-9e84-1a3e58ef6d3e-kube-api-access-jkh7h\") pod \"redhat-marketplace-wwp7w\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:03 crc kubenswrapper[4906]: I1123 07:03:03.836661 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:04 crc kubenswrapper[4906]: I1123 07:03:04.330941 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwp7w"] Nov 23 07:03:04 crc kubenswrapper[4906]: W1123 07:03:04.346123 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf47ba087_daee_498c_9e84_1a3e58ef6d3e.slice/crio-b85ac6cb620b84c4eb610a1a9987b463b0af9c453c32bdf5956b939f303a379d WatchSource:0}: Error finding container b85ac6cb620b84c4eb610a1a9987b463b0af9c453c32bdf5956b939f303a379d: Status 404 returned error can't find the container with id b85ac6cb620b84c4eb610a1a9987b463b0af9c453c32bdf5956b939f303a379d Nov 23 07:03:05 crc kubenswrapper[4906]: I1123 07:03:05.031518 4906 generic.go:334] "Generic (PLEG): container finished" podID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerID="8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49" exitCode=0 Nov 23 07:03:05 crc kubenswrapper[4906]: I1123 07:03:05.031648 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwp7w" event={"ID":"f47ba087-daee-498c-9e84-1a3e58ef6d3e","Type":"ContainerDied","Data":"8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49"} Nov 23 07:03:05 crc kubenswrapper[4906]: I1123 07:03:05.032235 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwp7w" event={"ID":"f47ba087-daee-498c-9e84-1a3e58ef6d3e","Type":"ContainerStarted","Data":"b85ac6cb620b84c4eb610a1a9987b463b0af9c453c32bdf5956b939f303a379d"} Nov 23 07:03:06 crc kubenswrapper[4906]: I1123 07:03:06.042988 4906 generic.go:334] "Generic (PLEG): container finished" podID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerID="247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267" exitCode=0 Nov 23 07:03:06 crc kubenswrapper[4906]: I1123 07:03:06.043054 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwp7w" event={"ID":"f47ba087-daee-498c-9e84-1a3e58ef6d3e","Type":"ContainerDied","Data":"247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267"} Nov 23 07:03:06 crc kubenswrapper[4906]: I1123 07:03:06.217648 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-tglkc" Nov 23 07:03:07 crc kubenswrapper[4906]: I1123 07:03:07.052236 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwp7w" event={"ID":"f47ba087-daee-498c-9e84-1a3e58ef6d3e","Type":"ContainerStarted","Data":"b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911"} Nov 23 07:03:07 crc kubenswrapper[4906]: I1123 07:03:07.076824 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wwp7w" podStartSLOduration=2.625274507 podStartE2EDuration="4.076799264s" podCreationTimestamp="2025-11-23 07:03:03 +0000 UTC" firstStartedPulling="2025-11-23 07:03:05.034602816 +0000 UTC m=+800.547994149" lastFinishedPulling="2025-11-23 07:03:06.486127593 +0000 UTC m=+801.999518906" observedRunningTime="2025-11-23 07:03:07.074221066 +0000 UTC m=+802.587612409" watchObservedRunningTime="2025-11-23 07:03:07.076799264 +0000 UTC m=+802.590190567" Nov 23 07:03:13 crc kubenswrapper[4906]: I1123 07:03:13.837947 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:13 crc kubenswrapper[4906]: I1123 07:03:13.839063 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:13 crc kubenswrapper[4906]: I1123 07:03:13.898609 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:14 crc kubenswrapper[4906]: I1123 07:03:14.154100 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:14 crc kubenswrapper[4906]: I1123 07:03:14.225142 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwp7w"] Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.122879 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wwp7w" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="registry-server" containerID="cri-o://b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911" gracePeriod=2 Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.536234 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.691311 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-utilities\") pod \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.691880 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkh7h\" (UniqueName: \"kubernetes.io/projected/f47ba087-daee-498c-9e84-1a3e58ef6d3e-kube-api-access-jkh7h\") pod \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.692043 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-catalog-content\") pod \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\" (UID: \"f47ba087-daee-498c-9e84-1a3e58ef6d3e\") " Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.693085 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-utilities" (OuterVolumeSpecName: "utilities") pod "f47ba087-daee-498c-9e84-1a3e58ef6d3e" (UID: "f47ba087-daee-498c-9e84-1a3e58ef6d3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.703981 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f47ba087-daee-498c-9e84-1a3e58ef6d3e-kube-api-access-jkh7h" (OuterVolumeSpecName: "kube-api-access-jkh7h") pod "f47ba087-daee-498c-9e84-1a3e58ef6d3e" (UID: "f47ba087-daee-498c-9e84-1a3e58ef6d3e"). InnerVolumeSpecName "kube-api-access-jkh7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.712433 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f47ba087-daee-498c-9e84-1a3e58ef6d3e" (UID: "f47ba087-daee-498c-9e84-1a3e58ef6d3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.793393 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.793443 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkh7h\" (UniqueName: \"kubernetes.io/projected/f47ba087-daee-498c-9e84-1a3e58ef6d3e-kube-api-access-jkh7h\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:16 crc kubenswrapper[4906]: I1123 07:03:16.793460 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f47ba087-daee-498c-9e84-1a3e58ef6d3e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.130360 4906 generic.go:334] "Generic (PLEG): container finished" podID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerID="b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911" exitCode=0 Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.130421 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwp7w" event={"ID":"f47ba087-daee-498c-9e84-1a3e58ef6d3e","Type":"ContainerDied","Data":"b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911"} Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.131726 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwp7w" event={"ID":"f47ba087-daee-498c-9e84-1a3e58ef6d3e","Type":"ContainerDied","Data":"b85ac6cb620b84c4eb610a1a9987b463b0af9c453c32bdf5956b939f303a379d"} Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.131750 4906 scope.go:117] "RemoveContainer" containerID="b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.130493 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwp7w" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.166967 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwp7w"] Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.173951 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwp7w"] Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.174350 4906 scope.go:117] "RemoveContainer" containerID="247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.192104 4906 scope.go:117] "RemoveContainer" containerID="8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.215457 4906 scope.go:117] "RemoveContainer" containerID="b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911" Nov 23 07:03:17 crc kubenswrapper[4906]: E1123 07:03:17.215990 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911\": container with ID starting with b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911 not found: ID does not exist" containerID="b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.216050 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911"} err="failed to get container status \"b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911\": rpc error: code = NotFound desc = could not find container \"b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911\": container with ID starting with b7a5656c0b674a014f6772f3bb36a7515aa5ed10e54ae6dc93ff4214ecd06911 not found: ID does not exist" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.216090 4906 scope.go:117] "RemoveContainer" containerID="247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267" Nov 23 07:03:17 crc kubenswrapper[4906]: E1123 07:03:17.216499 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267\": container with ID starting with 247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267 not found: ID does not exist" containerID="247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.216560 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267"} err="failed to get container status \"247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267\": rpc error: code = NotFound desc = could not find container \"247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267\": container with ID starting with 247caeec2897769cb6ec059a79300155b692a320a2d0e5e2fd9fbce25bef8267 not found: ID does not exist" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.216575 4906 scope.go:117] "RemoveContainer" containerID="8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49" Nov 23 07:03:17 crc kubenswrapper[4906]: E1123 07:03:17.217010 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49\": container with ID starting with 8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49 not found: ID does not exist" containerID="8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.217044 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49"} err="failed to get container status \"8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49\": rpc error: code = NotFound desc = could not find container \"8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49\": container with ID starting with 8a79d6c1a7a8503cce83c6aca50027d4b9564de6b361e73bd74dc890f4608f49 not found: ID does not exist" Nov 23 07:03:17 crc kubenswrapper[4906]: I1123 07:03:17.411757 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" path="/var/lib/kubelet/pods/f47ba087-daee-498c-9e84-1a3e58ef6d3e/volumes" Nov 23 07:03:20 crc kubenswrapper[4906]: I1123 07:03:20.945488 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:03:20 crc kubenswrapper[4906]: I1123 07:03:20.946104 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.067483 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-wp8fn" podUID="b7452b46-2b11-4a4f-893e-22bd995151bc" containerName="console" containerID="cri-o://6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79" gracePeriod=15 Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.100418 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8"] Nov 23 07:03:22 crc kubenswrapper[4906]: E1123 07:03:22.101071 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="registry-server" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.101090 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="registry-server" Nov 23 07:03:22 crc kubenswrapper[4906]: E1123 07:03:22.101101 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="extract-content" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.101109 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="extract-content" Nov 23 07:03:22 crc kubenswrapper[4906]: E1123 07:03:22.101125 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="extract-utilities" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.101135 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="extract-utilities" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.101287 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="f47ba087-daee-498c-9e84-1a3e58ef6d3e" containerName="registry-server" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.102201 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.104370 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.107348 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8"] Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.191074 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.191126 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7zrk\" (UniqueName: \"kubernetes.io/projected/2f94e5f1-58e2-4828-a11d-56daa984b926-kube-api-access-c7zrk\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.191156 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.293187 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.293246 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7zrk\" (UniqueName: \"kubernetes.io/projected/2f94e5f1-58e2-4828-a11d-56daa984b926-kube-api-access-c7zrk\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.293281 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.293941 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.294505 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.326808 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7zrk\" (UniqueName: \"kubernetes.io/projected/2f94e5f1-58e2-4828-a11d-56daa984b926-kube-api-access-c7zrk\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.456768 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wp8fn_b7452b46-2b11-4a4f-893e-22bd995151bc/console/0.log" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.456889 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.496547 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle\") pod \"b7452b46-2b11-4a4f-893e-22bd995151bc\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.496837 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config\") pod \"b7452b46-2b11-4a4f-893e-22bd995151bc\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.496903 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert\") pod \"b7452b46-2b11-4a4f-893e-22bd995151bc\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.497020 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgd69\" (UniqueName: \"kubernetes.io/projected/b7452b46-2b11-4a4f-893e-22bd995151bc-kube-api-access-lgd69\") pod \"b7452b46-2b11-4a4f-893e-22bd995151bc\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.497088 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config\") pod \"b7452b46-2b11-4a4f-893e-22bd995151bc\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.497154 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca\") pod \"b7452b46-2b11-4a4f-893e-22bd995151bc\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.497232 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert\") pod \"b7452b46-2b11-4a4f-893e-22bd995151bc\" (UID: \"b7452b46-2b11-4a4f-893e-22bd995151bc\") " Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.497604 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b7452b46-2b11-4a4f-893e-22bd995151bc" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.497733 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca" (OuterVolumeSpecName: "service-ca") pod "b7452b46-2b11-4a4f-893e-22bd995151bc" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.497825 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config" (OuterVolumeSpecName: "console-config") pod "b7452b46-2b11-4a4f-893e-22bd995151bc" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.498176 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b7452b46-2b11-4a4f-893e-22bd995151bc" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.499363 4906 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-console-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.499437 4906 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.499460 4906 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.499521 4906 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b7452b46-2b11-4a4f-893e-22bd995151bc-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.501913 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7452b46-2b11-4a4f-893e-22bd995151bc-kube-api-access-lgd69" (OuterVolumeSpecName: "kube-api-access-lgd69") pod "b7452b46-2b11-4a4f-893e-22bd995151bc" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc"). InnerVolumeSpecName "kube-api-access-lgd69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.502192 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b7452b46-2b11-4a4f-893e-22bd995151bc" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.503757 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b7452b46-2b11-4a4f-893e-22bd995151bc" (UID: "b7452b46-2b11-4a4f-893e-22bd995151bc"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.521720 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.609338 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgd69\" (UniqueName: \"kubernetes.io/projected/b7452b46-2b11-4a4f-893e-22bd995151bc-kube-api-access-lgd69\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.609974 4906 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.610006 4906 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b7452b46-2b11-4a4f-893e-22bd995151bc-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:22 crc kubenswrapper[4906]: I1123 07:03:22.794853 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8"] Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.174961 4906 generic.go:334] "Generic (PLEG): container finished" podID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerID="53e304b9c0e15791e9bf0f8d7eb23ae212e017546c70841928939f649c42a8fa" exitCode=0 Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.175026 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" event={"ID":"2f94e5f1-58e2-4828-a11d-56daa984b926","Type":"ContainerDied","Data":"53e304b9c0e15791e9bf0f8d7eb23ae212e017546c70841928939f649c42a8fa"} Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.175057 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" event={"ID":"2f94e5f1-58e2-4828-a11d-56daa984b926","Type":"ContainerStarted","Data":"5b8c0f89eab9616d7bf7d7f2eb755caa016f3a775025a6945c0f747f94208aa9"} Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.179319 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wp8fn_b7452b46-2b11-4a4f-893e-22bd995151bc/console/0.log" Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.179412 4906 generic.go:334] "Generic (PLEG): container finished" podID="b7452b46-2b11-4a4f-893e-22bd995151bc" containerID="6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79" exitCode=2 Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.179485 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wp8fn" event={"ID":"b7452b46-2b11-4a4f-893e-22bd995151bc","Type":"ContainerDied","Data":"6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79"} Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.179526 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wp8fn" event={"ID":"b7452b46-2b11-4a4f-893e-22bd995151bc","Type":"ContainerDied","Data":"0d3f3237ef143037775a6ebcbb81cd5e7f7f33915534c43237474f8c2c0a7e80"} Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.179549 4906 scope.go:117] "RemoveContainer" containerID="6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79" Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.179726 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wp8fn" Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.224626 4906 scope.go:117] "RemoveContainer" containerID="6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79" Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.225489 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wp8fn"] Nov 23 07:03:23 crc kubenswrapper[4906]: E1123 07:03:23.229558 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79\": container with ID starting with 6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79 not found: ID does not exist" containerID="6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79" Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.229609 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79"} err="failed to get container status \"6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79\": rpc error: code = NotFound desc = could not find container \"6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79\": container with ID starting with 6f6df7dc51261bf1f42f2c6eb9332e972bce7d47e7b84a1bae27b91a1e55bc79 not found: ID does not exist" Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.234362 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-wp8fn"] Nov 23 07:03:23 crc kubenswrapper[4906]: I1123 07:03:23.362885 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7452b46-2b11-4a4f-893e-22bd995151bc" path="/var/lib/kubelet/pods/b7452b46-2b11-4a4f-893e-22bd995151bc/volumes" Nov 23 07:03:25 crc kubenswrapper[4906]: I1123 07:03:25.203707 4906 generic.go:334] "Generic (PLEG): container finished" podID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerID="8f12b5c49f7af3f2c75d99f45b532a51fc1a6964e019666ec67ece1f69bc25ed" exitCode=0 Nov 23 07:03:25 crc kubenswrapper[4906]: I1123 07:03:25.203809 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" event={"ID":"2f94e5f1-58e2-4828-a11d-56daa984b926","Type":"ContainerDied","Data":"8f12b5c49f7af3f2c75d99f45b532a51fc1a6964e019666ec67ece1f69bc25ed"} Nov 23 07:03:26 crc kubenswrapper[4906]: I1123 07:03:26.217636 4906 generic.go:334] "Generic (PLEG): container finished" podID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerID="dc4edf633af7a220c038c89089c2008dc50de036e9a563ded03d7d8a85b46b78" exitCode=0 Nov 23 07:03:26 crc kubenswrapper[4906]: I1123 07:03:26.217716 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" event={"ID":"2f94e5f1-58e2-4828-a11d-56daa984b926","Type":"ContainerDied","Data":"dc4edf633af7a220c038c89089c2008dc50de036e9a563ded03d7d8a85b46b78"} Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.490412 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.584788 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-util\") pod \"2f94e5f1-58e2-4828-a11d-56daa984b926\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.585116 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-bundle\") pod \"2f94e5f1-58e2-4828-a11d-56daa984b926\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.585320 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7zrk\" (UniqueName: \"kubernetes.io/projected/2f94e5f1-58e2-4828-a11d-56daa984b926-kube-api-access-c7zrk\") pod \"2f94e5f1-58e2-4828-a11d-56daa984b926\" (UID: \"2f94e5f1-58e2-4828-a11d-56daa984b926\") " Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.586960 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-bundle" (OuterVolumeSpecName: "bundle") pod "2f94e5f1-58e2-4828-a11d-56daa984b926" (UID: "2f94e5f1-58e2-4828-a11d-56daa984b926"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.591417 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f94e5f1-58e2-4828-a11d-56daa984b926-kube-api-access-c7zrk" (OuterVolumeSpecName: "kube-api-access-c7zrk") pod "2f94e5f1-58e2-4828-a11d-56daa984b926" (UID: "2f94e5f1-58e2-4828-a11d-56daa984b926"). InnerVolumeSpecName "kube-api-access-c7zrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.617662 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-util" (OuterVolumeSpecName: "util") pod "2f94e5f1-58e2-4828-a11d-56daa984b926" (UID: "2f94e5f1-58e2-4828-a11d-56daa984b926"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.687518 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7zrk\" (UniqueName: \"kubernetes.io/projected/2f94e5f1-58e2-4828-a11d-56daa984b926-kube-api-access-c7zrk\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.687824 4906 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-util\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:27 crc kubenswrapper[4906]: I1123 07:03:27.687943 4906 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2f94e5f1-58e2-4828-a11d-56daa984b926-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:28 crc kubenswrapper[4906]: I1123 07:03:28.234952 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" event={"ID":"2f94e5f1-58e2-4828-a11d-56daa984b926","Type":"ContainerDied","Data":"5b8c0f89eab9616d7bf7d7f2eb755caa016f3a775025a6945c0f747f94208aa9"} Nov 23 07:03:28 crc kubenswrapper[4906]: I1123 07:03:28.235324 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b8c0f89eab9616d7bf7d7f2eb755caa016f3a775025a6945c0f747f94208aa9" Nov 23 07:03:28 crc kubenswrapper[4906]: I1123 07:03:28.235068 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.639365 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kqcqm"] Nov 23 07:03:29 crc kubenswrapper[4906]: E1123 07:03:29.639915 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7452b46-2b11-4a4f-893e-22bd995151bc" containerName="console" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.639929 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7452b46-2b11-4a4f-893e-22bd995151bc" containerName="console" Nov 23 07:03:29 crc kubenswrapper[4906]: E1123 07:03:29.639939 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerName="extract" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.639945 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerName="extract" Nov 23 07:03:29 crc kubenswrapper[4906]: E1123 07:03:29.639963 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerName="util" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.639971 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerName="util" Nov 23 07:03:29 crc kubenswrapper[4906]: E1123 07:03:29.639981 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerName="pull" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.639987 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerName="pull" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.640088 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7452b46-2b11-4a4f-893e-22bd995151bc" containerName="console" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.640102 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f94e5f1-58e2-4828-a11d-56daa984b926" containerName="extract" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.640919 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.658327 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqcqm"] Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.721100 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6pcb\" (UniqueName: \"kubernetes.io/projected/243d2e0a-a307-4077-9713-f408ebaae0a4-kube-api-access-b6pcb\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.721177 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-utilities\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.721210 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-catalog-content\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.822849 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-utilities\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.822930 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-catalog-content\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.823010 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6pcb\" (UniqueName: \"kubernetes.io/projected/243d2e0a-a307-4077-9713-f408ebaae0a4-kube-api-access-b6pcb\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.823887 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-utilities\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.823920 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-catalog-content\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.849304 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6pcb\" (UniqueName: \"kubernetes.io/projected/243d2e0a-a307-4077-9713-f408ebaae0a4-kube-api-access-b6pcb\") pod \"certified-operators-kqcqm\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:29 crc kubenswrapper[4906]: I1123 07:03:29.955271 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:30 crc kubenswrapper[4906]: I1123 07:03:30.215464 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqcqm"] Nov 23 07:03:30 crc kubenswrapper[4906]: I1123 07:03:30.249432 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqcqm" event={"ID":"243d2e0a-a307-4077-9713-f408ebaae0a4","Type":"ContainerStarted","Data":"304b3766937f0183b4cbe3cdb24c3fe8ad50f73be89ef9dcb0b937532d866ca1"} Nov 23 07:03:31 crc kubenswrapper[4906]: I1123 07:03:31.275703 4906 generic.go:334] "Generic (PLEG): container finished" podID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerID="f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0" exitCode=0 Nov 23 07:03:31 crc kubenswrapper[4906]: I1123 07:03:31.275816 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqcqm" event={"ID":"243d2e0a-a307-4077-9713-f408ebaae0a4","Type":"ContainerDied","Data":"f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0"} Nov 23 07:03:32 crc kubenswrapper[4906]: I1123 07:03:32.284104 4906 generic.go:334] "Generic (PLEG): container finished" podID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerID="ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104" exitCode=0 Nov 23 07:03:32 crc kubenswrapper[4906]: I1123 07:03:32.284179 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqcqm" event={"ID":"243d2e0a-a307-4077-9713-f408ebaae0a4","Type":"ContainerDied","Data":"ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104"} Nov 23 07:03:33 crc kubenswrapper[4906]: I1123 07:03:33.291811 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqcqm" event={"ID":"243d2e0a-a307-4077-9713-f408ebaae0a4","Type":"ContainerStarted","Data":"a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f"} Nov 23 07:03:33 crc kubenswrapper[4906]: I1123 07:03:33.311930 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kqcqm" podStartSLOduration=2.913466289 podStartE2EDuration="4.311895898s" podCreationTimestamp="2025-11-23 07:03:29 +0000 UTC" firstStartedPulling="2025-11-23 07:03:31.278286784 +0000 UTC m=+826.791678107" lastFinishedPulling="2025-11-23 07:03:32.676716413 +0000 UTC m=+828.190107716" observedRunningTime="2025-11-23 07:03:33.309915966 +0000 UTC m=+828.823307259" watchObservedRunningTime="2025-11-23 07:03:33.311895898 +0000 UTC m=+828.825287251" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.283108 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht"] Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.285254 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.294164 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tsrlb" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.297925 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.300325 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.300597 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.300794 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.321412 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht"] Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.342799 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2f26030-1eab-4685-a398-009687188c87-webhook-cert\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.342904 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpc4v\" (UniqueName: \"kubernetes.io/projected/f2f26030-1eab-4685-a398-009687188c87-kube-api-access-fpc4v\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.342945 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2f26030-1eab-4685-a398-009687188c87-apiservice-cert\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.444102 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpc4v\" (UniqueName: \"kubernetes.io/projected/f2f26030-1eab-4685-a398-009687188c87-kube-api-access-fpc4v\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.444191 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2f26030-1eab-4685-a398-009687188c87-apiservice-cert\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.445468 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2f26030-1eab-4685-a398-009687188c87-webhook-cert\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.452748 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f2f26030-1eab-4685-a398-009687188c87-apiservice-cert\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.455351 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f2f26030-1eab-4685-a398-009687188c87-webhook-cert\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.473752 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpc4v\" (UniqueName: \"kubernetes.io/projected/f2f26030-1eab-4685-a398-009687188c87-kube-api-access-fpc4v\") pod \"metallb-operator-controller-manager-67454dfd94-m6rht\" (UID: \"f2f26030-1eab-4685-a398-009687188c87\") " pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.545478 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74"] Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.546242 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.550781 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.551019 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-gd9z4" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.552667 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.567423 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74"] Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.610397 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.648846 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/deee94a4-f4d0-46f0-9307-99266349749a-apiservice-cert\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.649404 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/deee94a4-f4d0-46f0-9307-99266349749a-webhook-cert\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.649442 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq8fg\" (UniqueName: \"kubernetes.io/projected/deee94a4-f4d0-46f0-9307-99266349749a-kube-api-access-bq8fg\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.750715 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/deee94a4-f4d0-46f0-9307-99266349749a-webhook-cert\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.750771 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq8fg\" (UniqueName: \"kubernetes.io/projected/deee94a4-f4d0-46f0-9307-99266349749a-kube-api-access-bq8fg\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.750810 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/deee94a4-f4d0-46f0-9307-99266349749a-apiservice-cert\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.756047 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/deee94a4-f4d0-46f0-9307-99266349749a-apiservice-cert\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.762087 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/deee94a4-f4d0-46f0-9307-99266349749a-webhook-cert\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.779221 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq8fg\" (UniqueName: \"kubernetes.io/projected/deee94a4-f4d0-46f0-9307-99266349749a-kube-api-access-bq8fg\") pod \"metallb-operator-webhook-server-5fb958c5d-vnq74\" (UID: \"deee94a4-f4d0-46f0-9307-99266349749a\") " pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.894010 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:36 crc kubenswrapper[4906]: I1123 07:03:36.901900 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht"] Nov 23 07:03:36 crc kubenswrapper[4906]: W1123 07:03:36.921047 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2f26030_1eab_4685_a398_009687188c87.slice/crio-99e1b8ce30c8df735e536585624f0a4cc2ba3a90ee9805c1e150f93587bc2d20 WatchSource:0}: Error finding container 99e1b8ce30c8df735e536585624f0a4cc2ba3a90ee9805c1e150f93587bc2d20: Status 404 returned error can't find the container with id 99e1b8ce30c8df735e536585624f0a4cc2ba3a90ee9805c1e150f93587bc2d20 Nov 23 07:03:37 crc kubenswrapper[4906]: I1123 07:03:37.119778 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74"] Nov 23 07:03:37 crc kubenswrapper[4906]: I1123 07:03:37.333077 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" event={"ID":"deee94a4-f4d0-46f0-9307-99266349749a","Type":"ContainerStarted","Data":"bae2d01fe039f8781ce5b0d08b5bcc0c0e77ed41274a153bc01960a03e2486f1"} Nov 23 07:03:37 crc kubenswrapper[4906]: I1123 07:03:37.334312 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" event={"ID":"f2f26030-1eab-4685-a398-009687188c87","Type":"ContainerStarted","Data":"99e1b8ce30c8df735e536585624f0a4cc2ba3a90ee9805c1e150f93587bc2d20"} Nov 23 07:03:39 crc kubenswrapper[4906]: I1123 07:03:39.955997 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:39 crc kubenswrapper[4906]: I1123 07:03:39.956604 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:40 crc kubenswrapper[4906]: I1123 07:03:40.000041 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:40 crc kubenswrapper[4906]: I1123 07:03:40.411365 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:41 crc kubenswrapper[4906]: I1123 07:03:41.234030 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqcqm"] Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.379340 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" event={"ID":"f2f26030-1eab-4685-a398-009687188c87","Type":"ContainerStarted","Data":"6d786ae4a8d22de1ec9eee51f67b07ec603d85c11eaec44d6b3c61b268ea6588"} Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.379758 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.381931 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" event={"ID":"deee94a4-f4d0-46f0-9307-99266349749a","Type":"ContainerStarted","Data":"bd5bd3fbcdb6d6f9a58227539e595aed4ddee720a7b1188291305451ce650cdc"} Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.382115 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kqcqm" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="registry-server" containerID="cri-o://a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f" gracePeriod=2 Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.408029 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" podStartSLOduration=1.427548893 podStartE2EDuration="6.408009479s" podCreationTimestamp="2025-11-23 07:03:36 +0000 UTC" firstStartedPulling="2025-11-23 07:03:36.92623824 +0000 UTC m=+832.439629543" lastFinishedPulling="2025-11-23 07:03:41.906698826 +0000 UTC m=+837.420090129" observedRunningTime="2025-11-23 07:03:42.402725671 +0000 UTC m=+837.916117004" watchObservedRunningTime="2025-11-23 07:03:42.408009479 +0000 UTC m=+837.921400782" Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.430913 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" podStartSLOduration=1.630660166 podStartE2EDuration="6.430894959s" podCreationTimestamp="2025-11-23 07:03:36 +0000 UTC" firstStartedPulling="2025-11-23 07:03:37.129474256 +0000 UTC m=+832.642865559" lastFinishedPulling="2025-11-23 07:03:41.929709049 +0000 UTC m=+837.443100352" observedRunningTime="2025-11-23 07:03:42.427160331 +0000 UTC m=+837.940551634" watchObservedRunningTime="2025-11-23 07:03:42.430894959 +0000 UTC m=+837.944286272" Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.836154 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.942520 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-catalog-content\") pod \"243d2e0a-a307-4077-9713-f408ebaae0a4\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.942593 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-utilities\") pod \"243d2e0a-a307-4077-9713-f408ebaae0a4\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.942817 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6pcb\" (UniqueName: \"kubernetes.io/projected/243d2e0a-a307-4077-9713-f408ebaae0a4-kube-api-access-b6pcb\") pod \"243d2e0a-a307-4077-9713-f408ebaae0a4\" (UID: \"243d2e0a-a307-4077-9713-f408ebaae0a4\") " Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.943343 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-utilities" (OuterVolumeSpecName: "utilities") pod "243d2e0a-a307-4077-9713-f408ebaae0a4" (UID: "243d2e0a-a307-4077-9713-f408ebaae0a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.949750 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243d2e0a-a307-4077-9713-f408ebaae0a4-kube-api-access-b6pcb" (OuterVolumeSpecName: "kube-api-access-b6pcb") pod "243d2e0a-a307-4077-9713-f408ebaae0a4" (UID: "243d2e0a-a307-4077-9713-f408ebaae0a4"). InnerVolumeSpecName "kube-api-access-b6pcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:03:42 crc kubenswrapper[4906]: I1123 07:03:42.995956 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "243d2e0a-a307-4077-9713-f408ebaae0a4" (UID: "243d2e0a-a307-4077-9713-f408ebaae0a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.045912 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6pcb\" (UniqueName: \"kubernetes.io/projected/243d2e0a-a307-4077-9713-f408ebaae0a4-kube-api-access-b6pcb\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.045997 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.046016 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243d2e0a-a307-4077-9713-f408ebaae0a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.390485 4906 generic.go:334] "Generic (PLEG): container finished" podID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerID="a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f" exitCode=0 Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.391754 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqcqm" event={"ID":"243d2e0a-a307-4077-9713-f408ebaae0a4","Type":"ContainerDied","Data":"a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f"} Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.391821 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqcqm" event={"ID":"243d2e0a-a307-4077-9713-f408ebaae0a4","Type":"ContainerDied","Data":"304b3766937f0183b4cbe3cdb24c3fe8ad50f73be89ef9dcb0b937532d866ca1"} Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.391844 4906 scope.go:117] "RemoveContainer" containerID="a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.392423 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqcqm" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.392371 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.418575 4906 scope.go:117] "RemoveContainer" containerID="ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.439097 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqcqm"] Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.442511 4906 scope.go:117] "RemoveContainer" containerID="f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.444417 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kqcqm"] Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.482857 4906 scope.go:117] "RemoveContainer" containerID="a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f" Nov 23 07:03:43 crc kubenswrapper[4906]: E1123 07:03:43.483454 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f\": container with ID starting with a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f not found: ID does not exist" containerID="a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.483509 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f"} err="failed to get container status \"a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f\": rpc error: code = NotFound desc = could not find container \"a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f\": container with ID starting with a895b422daf17d534cc7207de0298845bc7ff8d638f929e24d5208760b0e7e2f not found: ID does not exist" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.483550 4906 scope.go:117] "RemoveContainer" containerID="ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104" Nov 23 07:03:43 crc kubenswrapper[4906]: E1123 07:03:43.484057 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104\": container with ID starting with ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104 not found: ID does not exist" containerID="ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.484102 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104"} err="failed to get container status \"ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104\": rpc error: code = NotFound desc = could not find container \"ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104\": container with ID starting with ec8628e2ee3ede3c711bfd45915906aa11def47f008aded9be2b434a81fd0104 not found: ID does not exist" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.484131 4906 scope.go:117] "RemoveContainer" containerID="f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0" Nov 23 07:03:43 crc kubenswrapper[4906]: E1123 07:03:43.484387 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0\": container with ID starting with f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0 not found: ID does not exist" containerID="f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0" Nov 23 07:03:43 crc kubenswrapper[4906]: I1123 07:03:43.484415 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0"} err="failed to get container status \"f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0\": rpc error: code = NotFound desc = could not find container \"f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0\": container with ID starting with f7ef4500f72d5d68f58c525a4694b0c378d4e774a0e8da77c6df99e6d8cbfad0 not found: ID does not exist" Nov 23 07:03:45 crc kubenswrapper[4906]: I1123 07:03:45.368297 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" path="/var/lib/kubelet/pods/243d2e0a-a307-4077-9713-f408ebaae0a4/volumes" Nov 23 07:03:50 crc kubenswrapper[4906]: I1123 07:03:50.945589 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:03:50 crc kubenswrapper[4906]: I1123 07:03:50.946302 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:03:50 crc kubenswrapper[4906]: I1123 07:03:50.946364 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:03:50 crc kubenswrapper[4906]: I1123 07:03:50.947180 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e4be889a6576c0c1dc1c3392a1a758863587cc0404504906531a0c6425bf526"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:03:50 crc kubenswrapper[4906]: I1123 07:03:50.947250 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://5e4be889a6576c0c1dc1c3392a1a758863587cc0404504906531a0c6425bf526" gracePeriod=600 Nov 23 07:03:51 crc kubenswrapper[4906]: I1123 07:03:51.463525 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="5e4be889a6576c0c1dc1c3392a1a758863587cc0404504906531a0c6425bf526" exitCode=0 Nov 23 07:03:51 crc kubenswrapper[4906]: I1123 07:03:51.463626 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"5e4be889a6576c0c1dc1c3392a1a758863587cc0404504906531a0c6425bf526"} Nov 23 07:03:51 crc kubenswrapper[4906]: I1123 07:03:51.464101 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"c632a9b1605581f702cb4b0b6cbc68235936754af5501ce83c90241ccf464cb9"} Nov 23 07:03:51 crc kubenswrapper[4906]: I1123 07:03:51.464134 4906 scope.go:117] "RemoveContainer" containerID="dd5ec0fa7f59d5550f15a2dfacce699aa4807dde1d3ef9fa91c46235d502f959" Nov 23 07:03:56 crc kubenswrapper[4906]: I1123 07:03:56.900784 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5fb958c5d-vnq74" Nov 23 07:04:16 crc kubenswrapper[4906]: I1123 07:04:16.613865 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-67454dfd94-m6rht" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.420023 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-9jk2v"] Nov 23 07:04:17 crc kubenswrapper[4906]: E1123 07:04:17.420506 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="extract-content" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.420532 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="extract-content" Nov 23 07:04:17 crc kubenswrapper[4906]: E1123 07:04:17.420576 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="registry-server" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.420590 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="registry-server" Nov 23 07:04:17 crc kubenswrapper[4906]: E1123 07:04:17.420603 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="extract-utilities" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.420613 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="extract-utilities" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.420798 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="243d2e0a-a307-4077-9713-f408ebaae0a4" containerName="registry-server" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.423577 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.423756 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-m45rz"] Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.424904 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.426222 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.428134 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-h59g8" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.428421 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.428593 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.443070 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-m45rz"] Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513596 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ad6a4f4d-bf8d-4628-8a89-7c67305e3b47-cert\") pod \"frr-k8s-webhook-server-6998585d5-m45rz\" (UID: \"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513698 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xvgz\" (UniqueName: \"kubernetes.io/projected/3bcd4535-f910-4cf2-8260-52d2c2506d7b-kube-api-access-8xvgz\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513774 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-metrics\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513797 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-reloader\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513819 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-sockets\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513841 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-startup\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513867 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bcd4535-f910-4cf2-8260-52d2c2506d7b-metrics-certs\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513953 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-conf\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.513984 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n857\" (UniqueName: \"kubernetes.io/projected/ad6a4f4d-bf8d-4628-8a89-7c67305e3b47-kube-api-access-4n857\") pod \"frr-k8s-webhook-server-6998585d5-m45rz\" (UID: \"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.541204 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hgss7"] Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.542163 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.544333 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.548593 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-gfx82" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.549115 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.554672 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.563781 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-vpptg"] Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.564754 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.573018 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.583290 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-vpptg"] Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.615087 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xvgz\" (UniqueName: \"kubernetes.io/projected/3bcd4535-f910-4cf2-8260-52d2c2506d7b-kube-api-access-8xvgz\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.615399 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-metrics-certs\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616134 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-metrics\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616242 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-reloader\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616350 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-sockets\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616446 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-startup\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616534 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bcd4535-f910-4cf2-8260-52d2c2506d7b-metrics-certs\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616619 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-conf\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616766 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n857\" (UniqueName: \"kubernetes.io/projected/ad6a4f4d-bf8d-4628-8a89-7c67305e3b47-kube-api-access-4n857\") pod \"frr-k8s-webhook-server-6998585d5-m45rz\" (UID: \"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616862 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a0c43812-7474-4ecc-a431-2f79a2441bf2-metallb-excludel2\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.616946 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5wkw\" (UniqueName: \"kubernetes.io/projected/a0c43812-7474-4ecc-a431-2f79a2441bf2-kube-api-access-j5wkw\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.617043 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ad6a4f4d-bf8d-4628-8a89-7c67305e3b47-cert\") pod \"frr-k8s-webhook-server-6998585d5-m45rz\" (UID: \"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.617118 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.618089 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-conf\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.618804 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-reloader\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.619042 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-sockets\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.619216 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3bcd4535-f910-4cf2-8260-52d2c2506d7b-frr-startup\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.619473 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3bcd4535-f910-4cf2-8260-52d2c2506d7b-metrics\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.625470 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bcd4535-f910-4cf2-8260-52d2c2506d7b-metrics-certs\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.626205 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ad6a4f4d-bf8d-4628-8a89-7c67305e3b47-cert\") pod \"frr-k8s-webhook-server-6998585d5-m45rz\" (UID: \"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.652461 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xvgz\" (UniqueName: \"kubernetes.io/projected/3bcd4535-f910-4cf2-8260-52d2c2506d7b-kube-api-access-8xvgz\") pod \"frr-k8s-9jk2v\" (UID: \"3bcd4535-f910-4cf2-8260-52d2c2506d7b\") " pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.663920 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n857\" (UniqueName: \"kubernetes.io/projected/ad6a4f4d-bf8d-4628-8a89-7c67305e3b47-kube-api-access-4n857\") pod \"frr-k8s-webhook-server-6998585d5-m45rz\" (UID: \"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.719324 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-cert\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.720425 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a0c43812-7474-4ecc-a431-2f79a2441bf2-metallb-excludel2\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.720492 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a0c43812-7474-4ecc-a431-2f79a2441bf2-metallb-excludel2\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.720537 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsrb5\" (UniqueName: \"kubernetes.io/projected/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-kube-api-access-nsrb5\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.720663 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5wkw\" (UniqueName: \"kubernetes.io/projected/a0c43812-7474-4ecc-a431-2f79a2441bf2-kube-api-access-j5wkw\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.720729 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-metrics-certs\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.720827 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.720949 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-metrics-certs\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: E1123 07:04:17.721136 4906 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 23 07:04:17 crc kubenswrapper[4906]: E1123 07:04:17.721219 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist podName:a0c43812-7474-4ecc-a431-2f79a2441bf2 nodeName:}" failed. No retries permitted until 2025-11-23 07:04:18.221196752 +0000 UTC m=+873.734588055 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist") pod "speaker-hgss7" (UID: "a0c43812-7474-4ecc-a431-2f79a2441bf2") : secret "metallb-memberlist" not found Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.724582 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-metrics-certs\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.736296 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5wkw\" (UniqueName: \"kubernetes.io/projected/a0c43812-7474-4ecc-a431-2f79a2441bf2-kube-api-access-j5wkw\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.752027 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.758109 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.822370 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-cert\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.822435 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsrb5\" (UniqueName: \"kubernetes.io/projected/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-kube-api-access-nsrb5\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.822468 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-metrics-certs\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.825756 4906 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.828791 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-metrics-certs\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.841213 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsrb5\" (UniqueName: \"kubernetes.io/projected/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-kube-api-access-nsrb5\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.841545 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/86aa5159-cbbd-44d3-88e0-cfd0c72e4592-cert\") pod \"controller-6c7b4b5f48-vpptg\" (UID: \"86aa5159-cbbd-44d3-88e0-cfd0c72e4592\") " pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:17 crc kubenswrapper[4906]: I1123 07:04:17.899386 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.014370 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-m45rz"] Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.200257 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-vpptg"] Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.240178 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:18 crc kubenswrapper[4906]: E1123 07:04:18.240400 4906 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 23 07:04:18 crc kubenswrapper[4906]: E1123 07:04:18.240862 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist podName:a0c43812-7474-4ecc-a431-2f79a2441bf2 nodeName:}" failed. No retries permitted until 2025-11-23 07:04:19.240833328 +0000 UTC m=+874.754224631 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist") pod "speaker-hgss7" (UID: "a0c43812-7474-4ecc-a431-2f79a2441bf2") : secret "metallb-memberlist" not found Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.663357 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-vpptg" event={"ID":"86aa5159-cbbd-44d3-88e0-cfd0c72e4592","Type":"ContainerStarted","Data":"a83fe25d2bf3d6e1707822600065a06aa725067b4f73dfa78c83d834f5c31b81"} Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.663406 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-vpptg" event={"ID":"86aa5159-cbbd-44d3-88e0-cfd0c72e4592","Type":"ContainerStarted","Data":"22ccb8e0ce1f3c97dd23ce517f12dedbc6e905212601f9282aed2f2bf8f59dc6"} Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.663415 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-vpptg" event={"ID":"86aa5159-cbbd-44d3-88e0-cfd0c72e4592","Type":"ContainerStarted","Data":"cd642ab88c195b7b3d8b5d37ce1a2015df974d7893a72af8276729afc7a52cec"} Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.664616 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.666585 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" event={"ID":"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47","Type":"ContainerStarted","Data":"30f4e39da0ac6f15075acd40363691b114e0a90986215700e95cf98b8bdfb9b8"} Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.668180 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerStarted","Data":"f42e1c1d300e27c67a0d004ca671a9269bfbfc4f273bdc953a38e11d81ad1da2"} Nov 23 07:04:18 crc kubenswrapper[4906]: I1123 07:04:18.689178 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-vpptg" podStartSLOduration=1.689151681 podStartE2EDuration="1.689151681s" podCreationTimestamp="2025-11-23 07:04:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:04:18.683280067 +0000 UTC m=+874.196671400" watchObservedRunningTime="2025-11-23 07:04:18.689151681 +0000 UTC m=+874.202543014" Nov 23 07:04:19 crc kubenswrapper[4906]: I1123 07:04:19.256390 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:19 crc kubenswrapper[4906]: I1123 07:04:19.264890 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a0c43812-7474-4ecc-a431-2f79a2441bf2-memberlist\") pod \"speaker-hgss7\" (UID: \"a0c43812-7474-4ecc-a431-2f79a2441bf2\") " pod="metallb-system/speaker-hgss7" Nov 23 07:04:19 crc kubenswrapper[4906]: I1123 07:04:19.356624 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hgss7" Nov 23 07:04:19 crc kubenswrapper[4906]: W1123 07:04:19.408084 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0c43812_7474_4ecc_a431_2f79a2441bf2.slice/crio-b0793b9e62a21b7e30f0ffe41b74861a924abef968d26be073dd6247f1ffe1be WatchSource:0}: Error finding container b0793b9e62a21b7e30f0ffe41b74861a924abef968d26be073dd6247f1ffe1be: Status 404 returned error can't find the container with id b0793b9e62a21b7e30f0ffe41b74861a924abef968d26be073dd6247f1ffe1be Nov 23 07:04:19 crc kubenswrapper[4906]: I1123 07:04:19.685042 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hgss7" event={"ID":"a0c43812-7474-4ecc-a431-2f79a2441bf2","Type":"ContainerStarted","Data":"b0793b9e62a21b7e30f0ffe41b74861a924abef968d26be073dd6247f1ffe1be"} Nov 23 07:04:20 crc kubenswrapper[4906]: I1123 07:04:20.694816 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hgss7" event={"ID":"a0c43812-7474-4ecc-a431-2f79a2441bf2","Type":"ContainerStarted","Data":"2862ec979b7b74c8ff9280277fc07cccf3ba90bf5eb7ac444859b5225754c827"} Nov 23 07:04:20 crc kubenswrapper[4906]: I1123 07:04:20.695289 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hgss7" event={"ID":"a0c43812-7474-4ecc-a431-2f79a2441bf2","Type":"ContainerStarted","Data":"0ced4435562e9d1f7ab952c3dc37adfd033180c3fd729fe5cd6d098af28798e1"} Nov 23 07:04:20 crc kubenswrapper[4906]: I1123 07:04:20.719515 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hgss7" podStartSLOduration=3.719484426 podStartE2EDuration="3.719484426s" podCreationTimestamp="2025-11-23 07:04:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:04:20.718060198 +0000 UTC m=+876.231451511" watchObservedRunningTime="2025-11-23 07:04:20.719484426 +0000 UTC m=+876.232875739" Nov 23 07:04:21 crc kubenswrapper[4906]: I1123 07:04:21.702396 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hgss7" Nov 23 07:04:26 crc kubenswrapper[4906]: I1123 07:04:26.754555 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" event={"ID":"ad6a4f4d-bf8d-4628-8a89-7c67305e3b47","Type":"ContainerStarted","Data":"968d956cb1bf0b782c46a423e2f132f4d78bcc532a4930b34336320903861065"} Nov 23 07:04:26 crc kubenswrapper[4906]: I1123 07:04:26.755588 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:26 crc kubenswrapper[4906]: I1123 07:04:26.758074 4906 generic.go:334] "Generic (PLEG): container finished" podID="3bcd4535-f910-4cf2-8260-52d2c2506d7b" containerID="a9442b1eb04356606dd08e63814ff43ab569641778329d7ad19f88d1ae6c7c42" exitCode=0 Nov 23 07:04:26 crc kubenswrapper[4906]: I1123 07:04:26.758113 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerDied","Data":"a9442b1eb04356606dd08e63814ff43ab569641778329d7ad19f88d1ae6c7c42"} Nov 23 07:04:26 crc kubenswrapper[4906]: I1123 07:04:26.781365 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" podStartSLOduration=1.536738948 podStartE2EDuration="9.781332325s" podCreationTimestamp="2025-11-23 07:04:17 +0000 UTC" firstStartedPulling="2025-11-23 07:04:18.028316357 +0000 UTC m=+873.541723251" lastFinishedPulling="2025-11-23 07:04:26.272925315 +0000 UTC m=+881.786316628" observedRunningTime="2025-11-23 07:04:26.776314104 +0000 UTC m=+882.289705437" watchObservedRunningTime="2025-11-23 07:04:26.781332325 +0000 UTC m=+882.294723678" Nov 23 07:04:27 crc kubenswrapper[4906]: I1123 07:04:27.770421 4906 generic.go:334] "Generic (PLEG): container finished" podID="3bcd4535-f910-4cf2-8260-52d2c2506d7b" containerID="81c9983a042d1f444d7d7c23b11e02fa65f6829a8381de25e42a8273fa7de960" exitCode=0 Nov 23 07:04:27 crc kubenswrapper[4906]: I1123 07:04:27.770560 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerDied","Data":"81c9983a042d1f444d7d7c23b11e02fa65f6829a8381de25e42a8273fa7de960"} Nov 23 07:04:28 crc kubenswrapper[4906]: I1123 07:04:28.796734 4906 generic.go:334] "Generic (PLEG): container finished" podID="3bcd4535-f910-4cf2-8260-52d2c2506d7b" containerID="28cb0ffdf27a6db6a80c98acffa1886a06e7fef865940672a720c43110b5bd52" exitCode=0 Nov 23 07:04:28 crc kubenswrapper[4906]: I1123 07:04:28.797172 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerDied","Data":"28cb0ffdf27a6db6a80c98acffa1886a06e7fef865940672a720c43110b5bd52"} Nov 23 07:04:29 crc kubenswrapper[4906]: I1123 07:04:29.368174 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hgss7" Nov 23 07:04:29 crc kubenswrapper[4906]: I1123 07:04:29.811587 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerStarted","Data":"06b0debd40650744796ca7c5b5b8cd798119890539bfb3df57d44fc5e1dc0b50"} Nov 23 07:04:29 crc kubenswrapper[4906]: I1123 07:04:29.811733 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerStarted","Data":"dc3ed787156b44ae8d70d1583af80d9dcec9139800fe3498958ede7095d5262f"} Nov 23 07:04:29 crc kubenswrapper[4906]: I1123 07:04:29.811767 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerStarted","Data":"7d584de1f3d4a926a0727d2103d542d49e2af499fb75910ec7b5dfc3e0905e23"} Nov 23 07:04:29 crc kubenswrapper[4906]: I1123 07:04:29.811794 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerStarted","Data":"ea9bcb05c39293d3cdda85605a7c6714115c99ca0c3a647e198b9a9e30738bb7"} Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.827564 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerStarted","Data":"ff1540bb38716f0a4fba8d8b3cbf1ddefabea2a7100ce45eb8215094310ec3ad"} Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.828092 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.828112 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9jk2v" event={"ID":"3bcd4535-f910-4cf2-8260-52d2c2506d7b","Type":"ContainerStarted","Data":"306aa6966f4de69918ef2292f9c5fe31bf60f8dc13a2f4c7fd0ce2584a52644f"} Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.869703 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-9jk2v" podStartSLOduration=5.490047469 podStartE2EDuration="13.869649882s" podCreationTimestamp="2025-11-23 07:04:17 +0000 UTC" firstStartedPulling="2025-11-23 07:04:17.918607286 +0000 UTC m=+873.431998589" lastFinishedPulling="2025-11-23 07:04:26.298209699 +0000 UTC m=+881.811601002" observedRunningTime="2025-11-23 07:04:30.866430248 +0000 UTC m=+886.379821561" watchObservedRunningTime="2025-11-23 07:04:30.869649882 +0000 UTC m=+886.383041225" Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.966021 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz"] Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.968794 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.973129 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 23 07:04:30 crc kubenswrapper[4906]: I1123 07:04:30.987810 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz"] Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.076840 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.077005 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh88g\" (UniqueName: \"kubernetes.io/projected/9a038721-ff13-448e-8ad9-59941a69a0f1-kube-api-access-mh88g\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.077045 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.179594 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.179813 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh88g\" (UniqueName: \"kubernetes.io/projected/9a038721-ff13-448e-8ad9-59941a69a0f1-kube-api-access-mh88g\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.179868 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.180807 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.180852 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.205458 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh88g\" (UniqueName: \"kubernetes.io/projected/9a038721-ff13-448e-8ad9-59941a69a0f1-kube-api-access-mh88g\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.286042 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.572030 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz"] Nov 23 07:04:31 crc kubenswrapper[4906]: W1123 07:04:31.587344 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a038721_ff13_448e_8ad9_59941a69a0f1.slice/crio-fa8e47c0e4e4f44e4578ec5ee1e1d0ab5fbb427810acbfd439d2a367fcc48ba8 WatchSource:0}: Error finding container fa8e47c0e4e4f44e4578ec5ee1e1d0ab5fbb427810acbfd439d2a367fcc48ba8: Status 404 returned error can't find the container with id fa8e47c0e4e4f44e4578ec5ee1e1d0ab5fbb427810acbfd439d2a367fcc48ba8 Nov 23 07:04:31 crc kubenswrapper[4906]: I1123 07:04:31.837908 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" event={"ID":"9a038721-ff13-448e-8ad9-59941a69a0f1","Type":"ContainerStarted","Data":"fa8e47c0e4e4f44e4578ec5ee1e1d0ab5fbb427810acbfd439d2a367fcc48ba8"} Nov 23 07:04:32 crc kubenswrapper[4906]: I1123 07:04:32.752704 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:32 crc kubenswrapper[4906]: I1123 07:04:32.845754 4906 generic.go:334] "Generic (PLEG): container finished" podID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerID="e1e8cf30d8b22333fb33ed536675125bd084ce5b1401c07b31d3115c991af255" exitCode=0 Nov 23 07:04:32 crc kubenswrapper[4906]: I1123 07:04:32.846385 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" event={"ID":"9a038721-ff13-448e-8ad9-59941a69a0f1","Type":"ContainerDied","Data":"e1e8cf30d8b22333fb33ed536675125bd084ce5b1401c07b31d3115c991af255"} Nov 23 07:04:32 crc kubenswrapper[4906]: I1123 07:04:32.856271 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:36 crc kubenswrapper[4906]: I1123 07:04:36.884961 4906 generic.go:334] "Generic (PLEG): container finished" podID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerID="c6b574534ee621e7969b343593c9216e16a6f1fe4c839f1522d783d4c9304b1b" exitCode=0 Nov 23 07:04:36 crc kubenswrapper[4906]: I1123 07:04:36.885746 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" event={"ID":"9a038721-ff13-448e-8ad9-59941a69a0f1","Type":"ContainerDied","Data":"c6b574534ee621e7969b343593c9216e16a6f1fe4c839f1522d783d4c9304b1b"} Nov 23 07:04:37 crc kubenswrapper[4906]: I1123 07:04:37.762802 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-m45rz" Nov 23 07:04:37 crc kubenswrapper[4906]: I1123 07:04:37.895843 4906 generic.go:334] "Generic (PLEG): container finished" podID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerID="ba8bc1295530185a3c5b937756167948a4c8c7b535cc583212cebf2e05181ed8" exitCode=0 Nov 23 07:04:37 crc kubenswrapper[4906]: I1123 07:04:37.895902 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" event={"ID":"9a038721-ff13-448e-8ad9-59941a69a0f1","Type":"ContainerDied","Data":"ba8bc1295530185a3c5b937756167948a4c8c7b535cc583212cebf2e05181ed8"} Nov 23 07:04:37 crc kubenswrapper[4906]: I1123 07:04:37.902894 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-vpptg" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.256876 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.320705 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-util\") pod \"9a038721-ff13-448e-8ad9-59941a69a0f1\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.320804 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-bundle\") pod \"9a038721-ff13-448e-8ad9-59941a69a0f1\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.320844 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh88g\" (UniqueName: \"kubernetes.io/projected/9a038721-ff13-448e-8ad9-59941a69a0f1-kube-api-access-mh88g\") pod \"9a038721-ff13-448e-8ad9-59941a69a0f1\" (UID: \"9a038721-ff13-448e-8ad9-59941a69a0f1\") " Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.323799 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-bundle" (OuterVolumeSpecName: "bundle") pod "9a038721-ff13-448e-8ad9-59941a69a0f1" (UID: "9a038721-ff13-448e-8ad9-59941a69a0f1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.328790 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a038721-ff13-448e-8ad9-59941a69a0f1-kube-api-access-mh88g" (OuterVolumeSpecName: "kube-api-access-mh88g") pod "9a038721-ff13-448e-8ad9-59941a69a0f1" (UID: "9a038721-ff13-448e-8ad9-59941a69a0f1"). InnerVolumeSpecName "kube-api-access-mh88g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.341999 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-util" (OuterVolumeSpecName: "util") pod "9a038721-ff13-448e-8ad9-59941a69a0f1" (UID: "9a038721-ff13-448e-8ad9-59941a69a0f1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.422643 4906 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-util\") on node \"crc\" DevicePath \"\"" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.422685 4906 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a038721-ff13-448e-8ad9-59941a69a0f1-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.422696 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh88g\" (UniqueName: \"kubernetes.io/projected/9a038721-ff13-448e-8ad9-59941a69a0f1-kube-api-access-mh88g\") on node \"crc\" DevicePath \"\"" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.910175 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" event={"ID":"9a038721-ff13-448e-8ad9-59941a69a0f1","Type":"ContainerDied","Data":"fa8e47c0e4e4f44e4578ec5ee1e1d0ab5fbb427810acbfd439d2a367fcc48ba8"} Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.910232 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa8e47c0e4e4f44e4578ec5ee1e1d0ab5fbb427810acbfd439d2a367fcc48ba8" Nov 23 07:04:39 crc kubenswrapper[4906]: I1123 07:04:39.910254 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.232415 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq"] Nov 23 07:04:44 crc kubenswrapper[4906]: E1123 07:04:44.233364 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerName="pull" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.233383 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerName="pull" Nov 23 07:04:44 crc kubenswrapper[4906]: E1123 07:04:44.233393 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerName="extract" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.233403 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerName="extract" Nov 23 07:04:44 crc kubenswrapper[4906]: E1123 07:04:44.233418 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerName="util" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.233427 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerName="util" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.233569 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a038721-ff13-448e-8ad9-59941a69a0f1" containerName="extract" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.234176 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.237412 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.237888 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.240321 4906 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-2rlnv" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.261774 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq"] Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.298305 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/625f9a5e-47ee-4dc6-9e6e-13271d84ae43-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwxjq\" (UID: \"625f9a5e-47ee-4dc6-9e6e-13271d84ae43\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.298434 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mptmt\" (UniqueName: \"kubernetes.io/projected/625f9a5e-47ee-4dc6-9e6e-13271d84ae43-kube-api-access-mptmt\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwxjq\" (UID: \"625f9a5e-47ee-4dc6-9e6e-13271d84ae43\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.399615 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/625f9a5e-47ee-4dc6-9e6e-13271d84ae43-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwxjq\" (UID: \"625f9a5e-47ee-4dc6-9e6e-13271d84ae43\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.399958 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mptmt\" (UniqueName: \"kubernetes.io/projected/625f9a5e-47ee-4dc6-9e6e-13271d84ae43-kube-api-access-mptmt\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwxjq\" (UID: \"625f9a5e-47ee-4dc6-9e6e-13271d84ae43\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.400186 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/625f9a5e-47ee-4dc6-9e6e-13271d84ae43-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwxjq\" (UID: \"625f9a5e-47ee-4dc6-9e6e-13271d84ae43\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.419903 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mptmt\" (UniqueName: \"kubernetes.io/projected/625f9a5e-47ee-4dc6-9e6e-13271d84ae43-kube-api-access-mptmt\") pod \"cert-manager-operator-controller-manager-64cf6dff88-jwxjq\" (UID: \"625f9a5e-47ee-4dc6-9e6e-13271d84ae43\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:44 crc kubenswrapper[4906]: I1123 07:04:44.552579 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" Nov 23 07:04:45 crc kubenswrapper[4906]: I1123 07:04:45.057710 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq"] Nov 23 07:04:45 crc kubenswrapper[4906]: W1123 07:04:45.065195 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod625f9a5e_47ee_4dc6_9e6e_13271d84ae43.slice/crio-e32d609cc478beda5fce51e310576eab8b767c5b41c023449e052af1ad5cd2fe WatchSource:0}: Error finding container e32d609cc478beda5fce51e310576eab8b767c5b41c023449e052af1ad5cd2fe: Status 404 returned error can't find the container with id e32d609cc478beda5fce51e310576eab8b767c5b41c023449e052af1ad5cd2fe Nov 23 07:04:45 crc kubenswrapper[4906]: I1123 07:04:45.953313 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" event={"ID":"625f9a5e-47ee-4dc6-9e6e-13271d84ae43","Type":"ContainerStarted","Data":"e32d609cc478beda5fce51e310576eab8b767c5b41c023449e052af1ad5cd2fe"} Nov 23 07:04:47 crc kubenswrapper[4906]: I1123 07:04:47.755578 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-9jk2v" Nov 23 07:04:53 crc kubenswrapper[4906]: I1123 07:04:53.003111 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" event={"ID":"625f9a5e-47ee-4dc6-9e6e-13271d84ae43","Type":"ContainerStarted","Data":"b4f614a1978a2c9d1f61f341606c6421d2bef77554a7bf145dd490638d434f1e"} Nov 23 07:04:53 crc kubenswrapper[4906]: I1123 07:04:53.038959 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-jwxjq" podStartSLOduration=2.06266515 podStartE2EDuration="9.038926163s" podCreationTimestamp="2025-11-23 07:04:44 +0000 UTC" firstStartedPulling="2025-11-23 07:04:45.066910373 +0000 UTC m=+900.580301676" lastFinishedPulling="2025-11-23 07:04:52.043171386 +0000 UTC m=+907.556562689" observedRunningTime="2025-11-23 07:04:53.03459597 +0000 UTC m=+908.547987313" watchObservedRunningTime="2025-11-23 07:04:53.038926163 +0000 UTC m=+908.552317506" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.704149 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-p4rqc"] Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.705350 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.708289 4906 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-8llmx" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.708700 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.709801 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.723851 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-p4rqc"] Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.844551 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a8e7a6e3-86d5-4d34-88d7-6fd220435b4e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-p4rqc\" (UID: \"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.845028 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9bsw\" (UniqueName: \"kubernetes.io/projected/a8e7a6e3-86d5-4d34-88d7-6fd220435b4e-kube-api-access-n9bsw\") pod \"cert-manager-webhook-f4fb5df64-p4rqc\" (UID: \"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.946938 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9bsw\" (UniqueName: \"kubernetes.io/projected/a8e7a6e3-86d5-4d34-88d7-6fd220435b4e-kube-api-access-n9bsw\") pod \"cert-manager-webhook-f4fb5df64-p4rqc\" (UID: \"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.947328 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a8e7a6e3-86d5-4d34-88d7-6fd220435b4e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-p4rqc\" (UID: \"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.976236 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a8e7a6e3-86d5-4d34-88d7-6fd220435b4e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-p4rqc\" (UID: \"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:57 crc kubenswrapper[4906]: I1123 07:04:57.977635 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9bsw\" (UniqueName: \"kubernetes.io/projected/a8e7a6e3-86d5-4d34-88d7-6fd220435b4e-kube-api-access-n9bsw\") pod \"cert-manager-webhook-f4fb5df64-p4rqc\" (UID: \"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:58 crc kubenswrapper[4906]: I1123 07:04:58.022260 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:04:58 crc kubenswrapper[4906]: I1123 07:04:58.494521 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-p4rqc"] Nov 23 07:04:58 crc kubenswrapper[4906]: W1123 07:04:58.510212 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8e7a6e3_86d5_4d34_88d7_6fd220435b4e.slice/crio-8972e079ea590c784fb5312e6f354469c998c389d169ae15dfc3e438b154588d WatchSource:0}: Error finding container 8972e079ea590c784fb5312e6f354469c998c389d169ae15dfc3e438b154588d: Status 404 returned error can't find the container with id 8972e079ea590c784fb5312e6f354469c998c389d169ae15dfc3e438b154588d Nov 23 07:04:59 crc kubenswrapper[4906]: I1123 07:04:59.045057 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" event={"ID":"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e","Type":"ContainerStarted","Data":"8972e079ea590c784fb5312e6f354469c998c389d169ae15dfc3e438b154588d"} Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.059700 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9"] Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.060544 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.064003 4906 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-f48jb" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.070546 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9"] Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.113744 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxjb\" (UniqueName: \"kubernetes.io/projected/b8ec1d86-a3b3-48eb-901d-8ddf495fa97e-kube-api-access-cpxjb\") pod \"cert-manager-cainjector-855d9ccff4-d8rz9\" (UID: \"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.113875 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b8ec1d86-a3b3-48eb-901d-8ddf495fa97e-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-d8rz9\" (UID: \"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.214857 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b8ec1d86-a3b3-48eb-901d-8ddf495fa97e-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-d8rz9\" (UID: \"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.214913 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxjb\" (UniqueName: \"kubernetes.io/projected/b8ec1d86-a3b3-48eb-901d-8ddf495fa97e-kube-api-access-cpxjb\") pod \"cert-manager-cainjector-855d9ccff4-d8rz9\" (UID: \"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.238787 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxjb\" (UniqueName: \"kubernetes.io/projected/b8ec1d86-a3b3-48eb-901d-8ddf495fa97e-kube-api-access-cpxjb\") pod \"cert-manager-cainjector-855d9ccff4-d8rz9\" (UID: \"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.240315 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b8ec1d86-a3b3-48eb-901d-8ddf495fa97e-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-d8rz9\" (UID: \"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.385450 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" Nov 23 07:05:01 crc kubenswrapper[4906]: I1123 07:05:01.624638 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9"] Nov 23 07:05:01 crc kubenswrapper[4906]: W1123 07:05:01.640655 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8ec1d86_a3b3_48eb_901d_8ddf495fa97e.slice/crio-1c6687e59a5fdce9f58fc0335838345610e0c7da830e37d7304ab2ce2a8f674b WatchSource:0}: Error finding container 1c6687e59a5fdce9f58fc0335838345610e0c7da830e37d7304ab2ce2a8f674b: Status 404 returned error can't find the container with id 1c6687e59a5fdce9f58fc0335838345610e0c7da830e37d7304ab2ce2a8f674b Nov 23 07:05:02 crc kubenswrapper[4906]: I1123 07:05:02.089556 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" event={"ID":"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e","Type":"ContainerStarted","Data":"1c6687e59a5fdce9f58fc0335838345610e0c7da830e37d7304ab2ce2a8f674b"} Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.623471 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qbrp4"] Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.624842 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.627451 4906 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jg46h" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.646949 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qbrp4"] Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.776805 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6-bound-sa-token\") pod \"cert-manager-86cb77c54b-qbrp4\" (UID: \"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6\") " pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.776861 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8pzd\" (UniqueName: \"kubernetes.io/projected/3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6-kube-api-access-r8pzd\") pod \"cert-manager-86cb77c54b-qbrp4\" (UID: \"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6\") " pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.878224 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6-bound-sa-token\") pod \"cert-manager-86cb77c54b-qbrp4\" (UID: \"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6\") " pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.878301 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8pzd\" (UniqueName: \"kubernetes.io/projected/3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6-kube-api-access-r8pzd\") pod \"cert-manager-86cb77c54b-qbrp4\" (UID: \"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6\") " pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.901606 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6-bound-sa-token\") pod \"cert-manager-86cb77c54b-qbrp4\" (UID: \"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6\") " pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.901808 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8pzd\" (UniqueName: \"kubernetes.io/projected/3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6-kube-api-access-r8pzd\") pod \"cert-manager-86cb77c54b-qbrp4\" (UID: \"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6\") " pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:04 crc kubenswrapper[4906]: I1123 07:05:04.956167 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-qbrp4" Nov 23 07:05:08 crc kubenswrapper[4906]: I1123 07:05:08.987112 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qbrp4"] Nov 23 07:05:08 crc kubenswrapper[4906]: W1123 07:05:08.996054 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3db5f0b3_e0b2_4cdf_aae1_39d42a31daa6.slice/crio-96c8010a375d4dd3c2dcc3d29d89122b9775f031ea1322c31699231ccf8fe2d3 WatchSource:0}: Error finding container 96c8010a375d4dd3c2dcc3d29d89122b9775f031ea1322c31699231ccf8fe2d3: Status 404 returned error can't find the container with id 96c8010a375d4dd3c2dcc3d29d89122b9775f031ea1322c31699231ccf8fe2d3 Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.157784 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" event={"ID":"b8ec1d86-a3b3-48eb-901d-8ddf495fa97e","Type":"ContainerStarted","Data":"0399c142c2059de5179ec90f439413e12912a3c1781119f5e6194d5c59378ec4"} Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.160523 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" event={"ID":"a8e7a6e3-86d5-4d34-88d7-6fd220435b4e","Type":"ContainerStarted","Data":"bd66300d0e8ef94e7bc8e3cabfe48228c94fd5e002ea1545f884eefb13b16d03"} Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.161031 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.162942 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-qbrp4" event={"ID":"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6","Type":"ContainerStarted","Data":"e61fbdada1885f052eb665daded6162ac72e73ce2285c66b9cc91b3f24234feb"} Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.163250 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-qbrp4" event={"ID":"3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6","Type":"ContainerStarted","Data":"96c8010a375d4dd3c2dcc3d29d89122b9775f031ea1322c31699231ccf8fe2d3"} Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.182157 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d8rz9" podStartSLOduration=0.982585778 podStartE2EDuration="8.182139213s" podCreationTimestamp="2025-11-23 07:05:01 +0000 UTC" firstStartedPulling="2025-11-23 07:05:01.644390287 +0000 UTC m=+917.157781590" lastFinishedPulling="2025-11-23 07:05:08.843943682 +0000 UTC m=+924.357335025" observedRunningTime="2025-11-23 07:05:09.178559489 +0000 UTC m=+924.691950792" watchObservedRunningTime="2025-11-23 07:05:09.182139213 +0000 UTC m=+924.695530516" Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.219637 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" podStartSLOduration=1.814456074 podStartE2EDuration="12.219612307s" podCreationTimestamp="2025-11-23 07:04:57 +0000 UTC" firstStartedPulling="2025-11-23 07:04:58.512862645 +0000 UTC m=+914.026253958" lastFinishedPulling="2025-11-23 07:05:08.918018888 +0000 UTC m=+924.431410191" observedRunningTime="2025-11-23 07:05:09.214924094 +0000 UTC m=+924.728315407" watchObservedRunningTime="2025-11-23 07:05:09.219612307 +0000 UTC m=+924.733003610" Nov 23 07:05:09 crc kubenswrapper[4906]: I1123 07:05:09.222724 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-qbrp4" podStartSLOduration=5.222711988 podStartE2EDuration="5.222711988s" podCreationTimestamp="2025-11-23 07:05:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:05:09.197880227 +0000 UTC m=+924.711271530" watchObservedRunningTime="2025-11-23 07:05:09.222711988 +0000 UTC m=+924.736103301" Nov 23 07:05:18 crc kubenswrapper[4906]: I1123 07:05:18.026296 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-p4rqc" Nov 23 07:05:31 crc kubenswrapper[4906]: I1123 07:05:31.895548 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pb6pq"] Nov 23 07:05:31 crc kubenswrapper[4906]: I1123 07:05:31.897926 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:31 crc kubenswrapper[4906]: I1123 07:05:31.913880 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-trb4s" Nov 23 07:05:31 crc kubenswrapper[4906]: I1123 07:05:31.914351 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 23 07:05:31 crc kubenswrapper[4906]: I1123 07:05:31.913903 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 23 07:05:31 crc kubenswrapper[4906]: I1123 07:05:31.930130 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pb6pq"] Nov 23 07:05:31 crc kubenswrapper[4906]: I1123 07:05:31.959805 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj2vf\" (UniqueName: \"kubernetes.io/projected/8013380e-43a3-47f5-bd1c-13d2b9525879-kube-api-access-fj2vf\") pod \"openstack-operator-index-pb6pq\" (UID: \"8013380e-43a3-47f5-bd1c-13d2b9525879\") " pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:32 crc kubenswrapper[4906]: I1123 07:05:32.078358 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj2vf\" (UniqueName: \"kubernetes.io/projected/8013380e-43a3-47f5-bd1c-13d2b9525879-kube-api-access-fj2vf\") pod \"openstack-operator-index-pb6pq\" (UID: \"8013380e-43a3-47f5-bd1c-13d2b9525879\") " pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:32 crc kubenswrapper[4906]: I1123 07:05:32.115719 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj2vf\" (UniqueName: \"kubernetes.io/projected/8013380e-43a3-47f5-bd1c-13d2b9525879-kube-api-access-fj2vf\") pod \"openstack-operator-index-pb6pq\" (UID: \"8013380e-43a3-47f5-bd1c-13d2b9525879\") " pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:32 crc kubenswrapper[4906]: I1123 07:05:32.279306 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:32 crc kubenswrapper[4906]: I1123 07:05:32.779995 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pb6pq"] Nov 23 07:05:33 crc kubenswrapper[4906]: I1123 07:05:33.387618 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pb6pq" event={"ID":"8013380e-43a3-47f5-bd1c-13d2b9525879","Type":"ContainerStarted","Data":"186d1b71d0e800fc9dfc1758f426bf76aa6f04cdf35f119c4fa1bd61b00f0c55"} Nov 23 07:05:34 crc kubenswrapper[4906]: I1123 07:05:34.401490 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pb6pq" event={"ID":"8013380e-43a3-47f5-bd1c-13d2b9525879","Type":"ContainerStarted","Data":"70bd0d60025ae564d2cb932cdc36f17b72f401ea7c54c75a8c16cd00ec1fbdcb"} Nov 23 07:05:34 crc kubenswrapper[4906]: I1123 07:05:34.432107 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pb6pq" podStartSLOduration=2.543327002 podStartE2EDuration="3.43207048s" podCreationTimestamp="2025-11-23 07:05:31 +0000 UTC" firstStartedPulling="2025-11-23 07:05:32.795548256 +0000 UTC m=+948.308939569" lastFinishedPulling="2025-11-23 07:05:33.684291734 +0000 UTC m=+949.197683047" observedRunningTime="2025-11-23 07:05:34.428056675 +0000 UTC m=+949.941448008" watchObservedRunningTime="2025-11-23 07:05:34.43207048 +0000 UTC m=+949.945461793" Nov 23 07:05:42 crc kubenswrapper[4906]: I1123 07:05:42.279577 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:42 crc kubenswrapper[4906]: I1123 07:05:42.280353 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:42 crc kubenswrapper[4906]: I1123 07:05:42.323509 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:42 crc kubenswrapper[4906]: I1123 07:05:42.527492 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-pb6pq" Nov 23 07:05:44 crc kubenswrapper[4906]: I1123 07:05:44.948663 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9"] Nov 23 07:05:44 crc kubenswrapper[4906]: I1123 07:05:44.951854 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:44 crc kubenswrapper[4906]: I1123 07:05:44.956046 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-4htxp" Nov 23 07:05:44 crc kubenswrapper[4906]: I1123 07:05:44.964154 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9"] Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.038389 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7klg\" (UniqueName: \"kubernetes.io/projected/785d6bed-d104-4bdf-85d4-efb0a6db387a-kube-api-access-l7klg\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.038543 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.038635 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.141070 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7klg\" (UniqueName: \"kubernetes.io/projected/785d6bed-d104-4bdf-85d4-efb0a6db387a-kube-api-access-l7klg\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.141249 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.141319 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.142049 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.145025 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.181808 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7klg\" (UniqueName: \"kubernetes.io/projected/785d6bed-d104-4bdf-85d4-efb0a6db387a-kube-api-access-l7klg\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.293391 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-4htxp" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.301051 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:45 crc kubenswrapper[4906]: I1123 07:05:45.853007 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9"] Nov 23 07:05:46 crc kubenswrapper[4906]: I1123 07:05:46.527511 4906 generic.go:334] "Generic (PLEG): container finished" podID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerID="23b387f626fd2351028ea89d2e6c3d383975552eadd801b7254280d52395633e" exitCode=0 Nov 23 07:05:46 crc kubenswrapper[4906]: I1123 07:05:46.527574 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" event={"ID":"785d6bed-d104-4bdf-85d4-efb0a6db387a","Type":"ContainerDied","Data":"23b387f626fd2351028ea89d2e6c3d383975552eadd801b7254280d52395633e"} Nov 23 07:05:46 crc kubenswrapper[4906]: I1123 07:05:46.527614 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" event={"ID":"785d6bed-d104-4bdf-85d4-efb0a6db387a","Type":"ContainerStarted","Data":"28a228142b146b0502f98a6601e81026a593a496c199c0b0438d2809476fcee1"} Nov 23 07:05:47 crc kubenswrapper[4906]: I1123 07:05:47.536511 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" event={"ID":"785d6bed-d104-4bdf-85d4-efb0a6db387a","Type":"ContainerStarted","Data":"974e936199d8b70b373f16fdabd196df305bd14139b2776e9a1201f3a5c4a3e0"} Nov 23 07:05:48 crc kubenswrapper[4906]: I1123 07:05:48.546658 4906 generic.go:334] "Generic (PLEG): container finished" podID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerID="974e936199d8b70b373f16fdabd196df305bd14139b2776e9a1201f3a5c4a3e0" exitCode=0 Nov 23 07:05:48 crc kubenswrapper[4906]: I1123 07:05:48.546839 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" event={"ID":"785d6bed-d104-4bdf-85d4-efb0a6db387a","Type":"ContainerDied","Data":"974e936199d8b70b373f16fdabd196df305bd14139b2776e9a1201f3a5c4a3e0"} Nov 23 07:05:49 crc kubenswrapper[4906]: I1123 07:05:49.562979 4906 generic.go:334] "Generic (PLEG): container finished" podID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerID="a1a6e98310ef59fea03193ce2efd27a675399aaeb3bad7b19b209a328b22f47d" exitCode=0 Nov 23 07:05:49 crc kubenswrapper[4906]: I1123 07:05:49.564437 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" event={"ID":"785d6bed-d104-4bdf-85d4-efb0a6db387a","Type":"ContainerDied","Data":"a1a6e98310ef59fea03193ce2efd27a675399aaeb3bad7b19b209a328b22f47d"} Nov 23 07:05:50 crc kubenswrapper[4906]: I1123 07:05:50.962823 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.052922 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-bundle\") pod \"785d6bed-d104-4bdf-85d4-efb0a6db387a\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.053059 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7klg\" (UniqueName: \"kubernetes.io/projected/785d6bed-d104-4bdf-85d4-efb0a6db387a-kube-api-access-l7klg\") pod \"785d6bed-d104-4bdf-85d4-efb0a6db387a\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.053105 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-util\") pod \"785d6bed-d104-4bdf-85d4-efb0a6db387a\" (UID: \"785d6bed-d104-4bdf-85d4-efb0a6db387a\") " Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.054438 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-bundle" (OuterVolumeSpecName: "bundle") pod "785d6bed-d104-4bdf-85d4-efb0a6db387a" (UID: "785d6bed-d104-4bdf-85d4-efb0a6db387a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.060633 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/785d6bed-d104-4bdf-85d4-efb0a6db387a-kube-api-access-l7klg" (OuterVolumeSpecName: "kube-api-access-l7klg") pod "785d6bed-d104-4bdf-85d4-efb0a6db387a" (UID: "785d6bed-d104-4bdf-85d4-efb0a6db387a"). InnerVolumeSpecName "kube-api-access-l7klg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.086857 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-util" (OuterVolumeSpecName: "util") pod "785d6bed-d104-4bdf-85d4-efb0a6db387a" (UID: "785d6bed-d104-4bdf-85d4-efb0a6db387a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.154764 4906 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.154808 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7klg\" (UniqueName: \"kubernetes.io/projected/785d6bed-d104-4bdf-85d4-efb0a6db387a-kube-api-access-l7klg\") on node \"crc\" DevicePath \"\"" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.154821 4906 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/785d6bed-d104-4bdf-85d4-efb0a6db387a-util\") on node \"crc\" DevicePath \"\"" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.583895 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" event={"ID":"785d6bed-d104-4bdf-85d4-efb0a6db387a","Type":"ContainerDied","Data":"28a228142b146b0502f98a6601e81026a593a496c199c0b0438d2809476fcee1"} Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.583943 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28a228142b146b0502f98a6601e81026a593a496c199c0b0438d2809476fcee1" Nov 23 07:05:51 crc kubenswrapper[4906]: I1123 07:05:51.584061 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.953774 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2"] Nov 23 07:05:57 crc kubenswrapper[4906]: E1123 07:05:57.954663 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerName="extract" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.954696 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerName="extract" Nov 23 07:05:57 crc kubenswrapper[4906]: E1123 07:05:57.954715 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerName="util" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.954724 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerName="util" Nov 23 07:05:57 crc kubenswrapper[4906]: E1123 07:05:57.954751 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerName="pull" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.954760 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerName="pull" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.954904 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="785d6bed-d104-4bdf-85d4-efb0a6db387a" containerName="extract" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.955656 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.960113 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-jj4qb" Nov 23 07:05:57 crc kubenswrapper[4906]: I1123 07:05:57.986326 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2"] Nov 23 07:05:58 crc kubenswrapper[4906]: I1123 07:05:58.066821 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6rl9\" (UniqueName: \"kubernetes.io/projected/57a5923c-45c0-4acd-ae5f-3920aa1c532f-kube-api-access-p6rl9\") pod \"openstack-operator-controller-operator-8486c7f98b-v78j2\" (UID: \"57a5923c-45c0-4acd-ae5f-3920aa1c532f\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" Nov 23 07:05:58 crc kubenswrapper[4906]: I1123 07:05:58.168035 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6rl9\" (UniqueName: \"kubernetes.io/projected/57a5923c-45c0-4acd-ae5f-3920aa1c532f-kube-api-access-p6rl9\") pod \"openstack-operator-controller-operator-8486c7f98b-v78j2\" (UID: \"57a5923c-45c0-4acd-ae5f-3920aa1c532f\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" Nov 23 07:05:58 crc kubenswrapper[4906]: I1123 07:05:58.194869 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6rl9\" (UniqueName: \"kubernetes.io/projected/57a5923c-45c0-4acd-ae5f-3920aa1c532f-kube-api-access-p6rl9\") pod \"openstack-operator-controller-operator-8486c7f98b-v78j2\" (UID: \"57a5923c-45c0-4acd-ae5f-3920aa1c532f\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" Nov 23 07:05:58 crc kubenswrapper[4906]: I1123 07:05:58.274188 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" Nov 23 07:05:58 crc kubenswrapper[4906]: I1123 07:05:58.495730 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2"] Nov 23 07:05:58 crc kubenswrapper[4906]: I1123 07:05:58.635414 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" event={"ID":"57a5923c-45c0-4acd-ae5f-3920aa1c532f","Type":"ContainerStarted","Data":"bbe96c8994924d1c664b73b5c1069e3c6522d9570df7e2d9a38231c7502aff6a"} Nov 23 07:06:03 crc kubenswrapper[4906]: I1123 07:06:03.670035 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" event={"ID":"57a5923c-45c0-4acd-ae5f-3920aa1c532f","Type":"ContainerStarted","Data":"429ac7693b0005546725aa9e766b9129f39afa89fbe073cc135b9dbdcbe14f28"} Nov 23 07:06:06 crc kubenswrapper[4906]: I1123 07:06:06.696584 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" event={"ID":"57a5923c-45c0-4acd-ae5f-3920aa1c532f","Type":"ContainerStarted","Data":"c7383f7eda1ae2a6d0c82b1636f221f9628184bb5dfca19e7e26ee9cbfee9edf"} Nov 23 07:06:06 crc kubenswrapper[4906]: I1123 07:06:06.697143 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" Nov 23 07:06:06 crc kubenswrapper[4906]: I1123 07:06:06.742330 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" podStartSLOduration=2.525258745 podStartE2EDuration="9.74230956s" podCreationTimestamp="2025-11-23 07:05:57 +0000 UTC" firstStartedPulling="2025-11-23 07:05:58.504404435 +0000 UTC m=+974.017795738" lastFinishedPulling="2025-11-23 07:06:05.72145524 +0000 UTC m=+981.234846553" observedRunningTime="2025-11-23 07:06:06.738390427 +0000 UTC m=+982.251781730" watchObservedRunningTime="2025-11-23 07:06:06.74230956 +0000 UTC m=+982.255700863" Nov 23 07:06:08 crc kubenswrapper[4906]: I1123 07:06:08.279504 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v78j2" Nov 23 07:06:20 crc kubenswrapper[4906]: I1123 07:06:20.946166 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:06:20 crc kubenswrapper[4906]: I1123 07:06:20.946977 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.921098 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv"] Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.923231 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.925653 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zf4t8" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.936489 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw"] Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.937877 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.940258 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qd4vn" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.944929 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg"] Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.946309 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.948853 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-jh2nr" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.952187 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49jrw\" (UniqueName: \"kubernetes.io/projected/f7a10077-2d8a-4b42-8683-bfc0f336f7d1-kube-api-access-49jrw\") pod \"barbican-operator-controller-manager-7768f8c84f-fkvbv\" (UID: \"f7a10077-2d8a-4b42-8683-bfc0f336f7d1\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.952237 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpz4q\" (UniqueName: \"kubernetes.io/projected/150266f3-2d08-4b22-b00e-a04fa8419820-kube-api-access-vpz4q\") pod \"cinder-operator-controller-manager-6d8fd67bf7-59wgw\" (UID: \"150266f3-2d08-4b22-b00e-a04fa8419820\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.952297 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccbxh\" (UniqueName: \"kubernetes.io/projected/3bc07001-7205-48f6-a348-e62df3c098da-kube-api-access-ccbxh\") pod \"designate-operator-controller-manager-56dfb6b67f-bgshg\" (UID: \"3bc07001-7205-48f6-a348-e62df3c098da\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.954605 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv"] Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.966219 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw"] Nov 23 07:06:41 crc kubenswrapper[4906]: I1123 07:06:41.975811 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.007787 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.040434 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.041605 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.041894 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.044985 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5v7d4" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.051138 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-zdxrd" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.054316 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfxxs\" (UniqueName: \"kubernetes.io/projected/e8480261-0abe-4b7d-b741-b76fe8126894-kube-api-access-nfxxs\") pod \"glance-operator-controller-manager-8667fbf6f6-jngn6\" (UID: \"e8480261-0abe-4b7d-b741-b76fe8126894\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.054383 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49jrw\" (UniqueName: \"kubernetes.io/projected/f7a10077-2d8a-4b42-8683-bfc0f336f7d1-kube-api-access-49jrw\") pod \"barbican-operator-controller-manager-7768f8c84f-fkvbv\" (UID: \"f7a10077-2d8a-4b42-8683-bfc0f336f7d1\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.054414 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpz4q\" (UniqueName: \"kubernetes.io/projected/150266f3-2d08-4b22-b00e-a04fa8419820-kube-api-access-vpz4q\") pod \"cinder-operator-controller-manager-6d8fd67bf7-59wgw\" (UID: \"150266f3-2d08-4b22-b00e-a04fa8419820\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.054453 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46hjw\" (UniqueName: \"kubernetes.io/projected/896605e2-1d27-4bda-a8bf-783a9fbd0294-kube-api-access-46hjw\") pod \"heat-operator-controller-manager-bf4c6585d-q6mmt\" (UID: \"896605e2-1d27-4bda-a8bf-783a9fbd0294\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.054473 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccbxh\" (UniqueName: \"kubernetes.io/projected/3bc07001-7205-48f6-a348-e62df3c098da-kube-api-access-ccbxh\") pod \"designate-operator-controller-manager-56dfb6b67f-bgshg\" (UID: \"3bc07001-7205-48f6-a348-e62df3c098da\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.082514 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.087624 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.088551 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccbxh\" (UniqueName: \"kubernetes.io/projected/3bc07001-7205-48f6-a348-e62df3c098da-kube-api-access-ccbxh\") pod \"designate-operator-controller-manager-56dfb6b67f-bgshg\" (UID: \"3bc07001-7205-48f6-a348-e62df3c098da\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.088969 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.094757 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.095958 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-5slnl" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.096151 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.098149 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49jrw\" (UniqueName: \"kubernetes.io/projected/f7a10077-2d8a-4b42-8683-bfc0f336f7d1-kube-api-access-49jrw\") pod \"barbican-operator-controller-manager-7768f8c84f-fkvbv\" (UID: \"f7a10077-2d8a-4b42-8683-bfc0f336f7d1\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.109778 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpz4q\" (UniqueName: \"kubernetes.io/projected/150266f3-2d08-4b22-b00e-a04fa8419820-kube-api-access-vpz4q\") pod \"cinder-operator-controller-manager-6d8fd67bf7-59wgw\" (UID: \"150266f3-2d08-4b22-b00e-a04fa8419820\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.119407 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.120429 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.124179 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-47w9g" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.156609 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfxxs\" (UniqueName: \"kubernetes.io/projected/e8480261-0abe-4b7d-b741-b76fe8126894-kube-api-access-nfxxs\") pod \"glance-operator-controller-manager-8667fbf6f6-jngn6\" (UID: \"e8480261-0abe-4b7d-b741-b76fe8126894\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.156717 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46hjw\" (UniqueName: \"kubernetes.io/projected/896605e2-1d27-4bda-a8bf-783a9fbd0294-kube-api-access-46hjw\") pod \"heat-operator-controller-manager-bf4c6585d-q6mmt\" (UID: \"896605e2-1d27-4bda-a8bf-783a9fbd0294\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.158638 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.176901 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfxxs\" (UniqueName: \"kubernetes.io/projected/e8480261-0abe-4b7d-b741-b76fe8126894-kube-api-access-nfxxs\") pod \"glance-operator-controller-manager-8667fbf6f6-jngn6\" (UID: \"e8480261-0abe-4b7d-b741-b76fe8126894\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.179996 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.185370 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46hjw\" (UniqueName: \"kubernetes.io/projected/896605e2-1d27-4bda-a8bf-783a9fbd0294-kube-api-access-46hjw\") pod \"heat-operator-controller-manager-bf4c6585d-q6mmt\" (UID: \"896605e2-1d27-4bda-a8bf-783a9fbd0294\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.196589 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.198257 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.205127 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.206564 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.206959 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vnbf9" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.208936 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hsccq" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.231558 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.232719 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.234803 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6prmm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.244011 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.252457 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.263652 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c9510f-29ed-482f-8b63-377479eff207-cert\") pod \"infra-operator-controller-manager-769d9c7585-hpwgw\" (UID: \"29c9510f-29ed-482f-8b63-377479eff207\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.267190 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm74m\" (UniqueName: \"kubernetes.io/projected/29c9510f-29ed-482f-8b63-377479eff207-kube-api-access-bm74m\") pod \"infra-operator-controller-manager-769d9c7585-hpwgw\" (UID: \"29c9510f-29ed-482f-8b63-377479eff207\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.268826 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nj86\" (UniqueName: \"kubernetes.io/projected/bb4b51f3-5f0e-4a41-920c-7a179c1e0032-kube-api-access-2nj86\") pod \"horizon-operator-controller-manager-5d86b44686-vqg2n\" (UID: \"bb4b51f3-5f0e-4a41-920c-7a179c1e0032\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.268862 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgwgc\" (UniqueName: \"kubernetes.io/projected/45a30ceb-58e3-4e36-96a0-b64f35fa5d5e-kube-api-access-xgwgc\") pod \"manila-operator-controller-manager-7bb88cb858-xm76f\" (UID: \"45a30ceb-58e3-4e36-96a0-b64f35fa5d5e\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.274203 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.274580 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.285356 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.304593 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.306420 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.308863 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-snjrr" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.310709 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.321347 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.323107 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.325815 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-vltwd" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.328124 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.329316 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.330954 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-b4hgf" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.338902 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.347156 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.349990 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.355963 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.360151 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.363884 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-f5pc4" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.367515 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.370917 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.371639 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374169 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g7mg\" (UniqueName: \"kubernetes.io/projected/3247edc2-e84e-4c6b-9b60-cbfb7a7ec442-kube-api-access-7g7mg\") pod \"keystone-operator-controller-manager-7879fb76fd-8tg5j\" (UID: \"3247edc2-e84e-4c6b-9b60-cbfb7a7ec442\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374205 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9g9j\" (UniqueName: \"kubernetes.io/projected/f98f97de-580b-460d-9576-1b496313001f-kube-api-access-v9g9j\") pod \"neutron-operator-controller-manager-66b7d6f598-7d26f\" (UID: \"f98f97de-580b-460d-9576-1b496313001f\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374262 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm74m\" (UniqueName: \"kubernetes.io/projected/29c9510f-29ed-482f-8b63-377479eff207-kube-api-access-bm74m\") pod \"infra-operator-controller-manager-769d9c7585-hpwgw\" (UID: \"29c9510f-29ed-482f-8b63-377479eff207\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374303 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzbrc\" (UniqueName: \"kubernetes.io/projected/b9bdc0d6-3a2e-4eda-8c73-870215a2018d-kube-api-access-nzbrc\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-c9xlv\" (UID: \"b9bdc0d6-3a2e-4eda-8c73-870215a2018d\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374329 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nj86\" (UniqueName: \"kubernetes.io/projected/bb4b51f3-5f0e-4a41-920c-7a179c1e0032-kube-api-access-2nj86\") pod \"horizon-operator-controller-manager-5d86b44686-vqg2n\" (UID: \"bb4b51f3-5f0e-4a41-920c-7a179c1e0032\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374352 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcs4b\" (UniqueName: \"kubernetes.io/projected/c4eaa84e-ff1f-4a9a-9ffe-56350431eb01-kube-api-access-tcs4b\") pod \"ironic-operator-controller-manager-5c75d7c94b-xvpjm\" (UID: \"c4eaa84e-ff1f-4a9a-9ffe-56350431eb01\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374372 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgwgc\" (UniqueName: \"kubernetes.io/projected/45a30ceb-58e3-4e36-96a0-b64f35fa5d5e-kube-api-access-xgwgc\") pod \"manila-operator-controller-manager-7bb88cb858-xm76f\" (UID: \"45a30ceb-58e3-4e36-96a0-b64f35fa5d5e\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374413 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljzwz\" (UniqueName: \"kubernetes.io/projected/962c14dd-6e7a-4927-ba5b-84384bde4312-kube-api-access-ljzwz\") pod \"nova-operator-controller-manager-86d796d84d-kqzb7\" (UID: \"962c14dd-6e7a-4927-ba5b-84384bde4312\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374452 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c9510f-29ed-482f-8b63-377479eff207-cert\") pod \"infra-operator-controller-manager-769d9c7585-hpwgw\" (UID: \"29c9510f-29ed-482f-8b63-377479eff207\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374461 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.374477 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkfkg\" (UniqueName: \"kubernetes.io/projected/cdd2bdf7-90ef-42f0-9395-b4889673d69c-kube-api-access-nkfkg\") pod \"octavia-operator-controller-manager-6fdc856c5d-q98m4\" (UID: \"cdd2bdf7-90ef-42f0-9395-b4889673d69c\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" Nov 23 07:06:42 crc kubenswrapper[4906]: E1123 07:06:42.375418 4906 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 23 07:06:42 crc kubenswrapper[4906]: E1123 07:06:42.375481 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29c9510f-29ed-482f-8b63-377479eff207-cert podName:29c9510f-29ed-482f-8b63-377479eff207 nodeName:}" failed. No retries permitted until 2025-11-23 07:06:42.875459477 +0000 UTC m=+1018.388850780 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29c9510f-29ed-482f-8b63-377479eff207-cert") pod "infra-operator-controller-manager-769d9c7585-hpwgw" (UID: "29c9510f-29ed-482f-8b63-377479eff207") : secret "infra-operator-webhook-server-cert" not found Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.380037 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-hkvn9" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.380975 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.382659 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.392945 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.393189 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rx7s2" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.399448 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nj86\" (UniqueName: \"kubernetes.io/projected/bb4b51f3-5f0e-4a41-920c-7a179c1e0032-kube-api-access-2nj86\") pod \"horizon-operator-controller-manager-5d86b44686-vqg2n\" (UID: \"bb4b51f3-5f0e-4a41-920c-7a179c1e0032\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.405155 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.414726 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgwgc\" (UniqueName: \"kubernetes.io/projected/45a30ceb-58e3-4e36-96a0-b64f35fa5d5e-kube-api-access-xgwgc\") pod \"manila-operator-controller-manager-7bb88cb858-xm76f\" (UID: \"45a30ceb-58e3-4e36-96a0-b64f35fa5d5e\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.420776 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.425231 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.448215 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm74m\" (UniqueName: \"kubernetes.io/projected/29c9510f-29ed-482f-8b63-377479eff207-kube-api-access-bm74m\") pod \"infra-operator-controller-manager-769d9c7585-hpwgw\" (UID: \"29c9510f-29ed-482f-8b63-377479eff207\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.451557 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.454960 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vsfx8" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.460187 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.469980 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.473362 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.484270 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljzwz\" (UniqueName: \"kubernetes.io/projected/962c14dd-6e7a-4927-ba5b-84384bde4312-kube-api-access-ljzwz\") pod \"nova-operator-controller-manager-86d796d84d-kqzb7\" (UID: \"962c14dd-6e7a-4927-ba5b-84384bde4312\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.484532 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbm4r\" (UniqueName: \"kubernetes.io/projected/dd441d38-8fef-4ee4-8e92-f32682d524c3-kube-api-access-xbm4r\") pod \"placement-operator-controller-manager-6dc664666c-ndc2b\" (UID: \"dd441d38-8fef-4ee4-8e92-f32682d524c3\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.484621 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d43ca54d-dde3-4178-bb66-1bf010c5dac8-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd4489m6x\" (UID: \"d43ca54d-dde3-4178-bb66-1bf010c5dac8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.484745 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkfkg\" (UniqueName: \"kubernetes.io/projected/cdd2bdf7-90ef-42f0-9395-b4889673d69c-kube-api-access-nkfkg\") pod \"octavia-operator-controller-manager-6fdc856c5d-q98m4\" (UID: \"cdd2bdf7-90ef-42f0-9395-b4889673d69c\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.484785 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvhl\" (UniqueName: \"kubernetes.io/projected/d43ca54d-dde3-4178-bb66-1bf010c5dac8-kube-api-access-lnvhl\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd4489m6x\" (UID: \"d43ca54d-dde3-4178-bb66-1bf010c5dac8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.484817 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g7mg\" (UniqueName: \"kubernetes.io/projected/3247edc2-e84e-4c6b-9b60-cbfb7a7ec442-kube-api-access-7g7mg\") pod \"keystone-operator-controller-manager-7879fb76fd-8tg5j\" (UID: \"3247edc2-e84e-4c6b-9b60-cbfb7a7ec442\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.484970 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9g9j\" (UniqueName: \"kubernetes.io/projected/f98f97de-580b-460d-9576-1b496313001f-kube-api-access-v9g9j\") pod \"neutron-operator-controller-manager-66b7d6f598-7d26f\" (UID: \"f98f97de-580b-460d-9576-1b496313001f\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.485151 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzbrc\" (UniqueName: \"kubernetes.io/projected/b9bdc0d6-3a2e-4eda-8c73-870215a2018d-kube-api-access-nzbrc\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-c9xlv\" (UID: \"b9bdc0d6-3a2e-4eda-8c73-870215a2018d\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.485186 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5qrj\" (UniqueName: \"kubernetes.io/projected/128c4cf7-46dc-4438-a13e-ebb14600b794-kube-api-access-l5qrj\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-gd2gm\" (UID: \"128c4cf7-46dc-4438-a13e-ebb14600b794\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.485218 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcs4b\" (UniqueName: \"kubernetes.io/projected/c4eaa84e-ff1f-4a9a-9ffe-56350431eb01-kube-api-access-tcs4b\") pod \"ironic-operator-controller-manager-5c75d7c94b-xvpjm\" (UID: \"c4eaa84e-ff1f-4a9a-9ffe-56350431eb01\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.491598 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.500244 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.510228 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-47qtd" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.517377 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.526077 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzbrc\" (UniqueName: \"kubernetes.io/projected/b9bdc0d6-3a2e-4eda-8c73-870215a2018d-kube-api-access-nzbrc\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-c9xlv\" (UID: \"b9bdc0d6-3a2e-4eda-8c73-870215a2018d\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.531916 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g7mg\" (UniqueName: \"kubernetes.io/projected/3247edc2-e84e-4c6b-9b60-cbfb7a7ec442-kube-api-access-7g7mg\") pod \"keystone-operator-controller-manager-7879fb76fd-8tg5j\" (UID: \"3247edc2-e84e-4c6b-9b60-cbfb7a7ec442\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.532367 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkfkg\" (UniqueName: \"kubernetes.io/projected/cdd2bdf7-90ef-42f0-9395-b4889673d69c-kube-api-access-nkfkg\") pod \"octavia-operator-controller-manager-6fdc856c5d-q98m4\" (UID: \"cdd2bdf7-90ef-42f0-9395-b4889673d69c\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.533835 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcs4b\" (UniqueName: \"kubernetes.io/projected/c4eaa84e-ff1f-4a9a-9ffe-56350431eb01-kube-api-access-tcs4b\") pod \"ironic-operator-controller-manager-5c75d7c94b-xvpjm\" (UID: \"c4eaa84e-ff1f-4a9a-9ffe-56350431eb01\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.534416 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9g9j\" (UniqueName: \"kubernetes.io/projected/f98f97de-580b-460d-9576-1b496313001f-kube-api-access-v9g9j\") pod \"neutron-operator-controller-manager-66b7d6f598-7d26f\" (UID: \"f98f97de-580b-460d-9576-1b496313001f\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.543802 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.544817 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljzwz\" (UniqueName: \"kubernetes.io/projected/962c14dd-6e7a-4927-ba5b-84384bde4312-kube-api-access-ljzwz\") pod \"nova-operator-controller-manager-86d796d84d-kqzb7\" (UID: \"962c14dd-6e7a-4927-ba5b-84384bde4312\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.556930 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.573922 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.574372 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.575581 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.580716 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jql59" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.586461 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.587282 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvhl\" (UniqueName: \"kubernetes.io/projected/d43ca54d-dde3-4178-bb66-1bf010c5dac8-kube-api-access-lnvhl\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd4489m6x\" (UID: \"d43ca54d-dde3-4178-bb66-1bf010c5dac8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.587356 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vz8x\" (UniqueName: \"kubernetes.io/projected/f050d900-ee68-4bdb-ad1f-3e64bf8db957-kube-api-access-8vz8x\") pod \"swift-operator-controller-manager-799cb6ffd6-wtlrm\" (UID: \"f050d900-ee68-4bdb-ad1f-3e64bf8db957\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.587397 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5qrj\" (UniqueName: \"kubernetes.io/projected/128c4cf7-46dc-4438-a13e-ebb14600b794-kube-api-access-l5qrj\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-gd2gm\" (UID: \"128c4cf7-46dc-4438-a13e-ebb14600b794\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.587435 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjwlz\" (UniqueName: \"kubernetes.io/projected/2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239-kube-api-access-fjwlz\") pod \"telemetry-operator-controller-manager-7798859c74-f4jtr\" (UID: \"2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.588398 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbm4r\" (UniqueName: \"kubernetes.io/projected/dd441d38-8fef-4ee4-8e92-f32682d524c3-kube-api-access-xbm4r\") pod \"placement-operator-controller-manager-6dc664666c-ndc2b\" (UID: \"dd441d38-8fef-4ee4-8e92-f32682d524c3\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.588459 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d43ca54d-dde3-4178-bb66-1bf010c5dac8-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd4489m6x\" (UID: \"d43ca54d-dde3-4178-bb66-1bf010c5dac8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:42 crc kubenswrapper[4906]: E1123 07:06:42.588599 4906 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 07:06:42 crc kubenswrapper[4906]: E1123 07:06:42.588655 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d43ca54d-dde3-4178-bb66-1bf010c5dac8-cert podName:d43ca54d-dde3-4178-bb66-1bf010c5dac8 nodeName:}" failed. No retries permitted until 2025-11-23 07:06:43.088639988 +0000 UTC m=+1018.602031291 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d43ca54d-dde3-4178-bb66-1bf010c5dac8-cert") pod "openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" (UID: "d43ca54d-dde3-4178-bb66-1bf010c5dac8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.610785 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbm4r\" (UniqueName: \"kubernetes.io/projected/dd441d38-8fef-4ee4-8e92-f32682d524c3-kube-api-access-xbm4r\") pod \"placement-operator-controller-manager-6dc664666c-ndc2b\" (UID: \"dd441d38-8fef-4ee4-8e92-f32682d524c3\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.619334 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5qrj\" (UniqueName: \"kubernetes.io/projected/128c4cf7-46dc-4438-a13e-ebb14600b794-kube-api-access-l5qrj\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-gd2gm\" (UID: \"128c4cf7-46dc-4438-a13e-ebb14600b794\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.622256 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvhl\" (UniqueName: \"kubernetes.io/projected/d43ca54d-dde3-4178-bb66-1bf010c5dac8-kube-api-access-lnvhl\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd4489m6x\" (UID: \"d43ca54d-dde3-4178-bb66-1bf010c5dac8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.649304 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.660177 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.671555 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.676087 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.680250 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.690634 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-s8stt" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.691192 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjwlz\" (UniqueName: \"kubernetes.io/projected/2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239-kube-api-access-fjwlz\") pod \"telemetry-operator-controller-manager-7798859c74-f4jtr\" (UID: \"2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.691652 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vz8x\" (UniqueName: \"kubernetes.io/projected/f050d900-ee68-4bdb-ad1f-3e64bf8db957-kube-api-access-8vz8x\") pod \"swift-operator-controller-manager-799cb6ffd6-wtlrm\" (UID: \"f050d900-ee68-4bdb-ad1f-3e64bf8db957\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.692073 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.692528 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.703006 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.718749 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vz8x\" (UniqueName: \"kubernetes.io/projected/f050d900-ee68-4bdb-ad1f-3e64bf8db957-kube-api-access-8vz8x\") pod \"swift-operator-controller-manager-799cb6ffd6-wtlrm\" (UID: \"f050d900-ee68-4bdb-ad1f-3e64bf8db957\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.723598 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.728933 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjwlz\" (UniqueName: \"kubernetes.io/projected/2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239-kube-api-access-fjwlz\") pod \"telemetry-operator-controller-manager-7798859c74-f4jtr\" (UID: \"2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.733538 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.737990 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-bxj4m" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.771365 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.793105 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4z9m\" (UniqueName: \"kubernetes.io/projected/c72780f9-1a11-4ddc-bf65-3b1eb726be9e-kube-api-access-q4z9m\") pod \"test-operator-controller-manager-8464cf66df-s4mqg\" (UID: \"c72780f9-1a11-4ddc-bf65-3b1eb726be9e\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.828820 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.831155 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.834174 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-z2blj" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.834421 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.848908 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.870340 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.875236 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.897054 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6n6d\" (UniqueName: \"kubernetes.io/projected/44c3b95a-a75b-4121-967a-28d483eb72a2-kube-api-access-x6n6d\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.897142 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c9510f-29ed-482f-8b63-377479eff207-cert\") pod \"infra-operator-controller-manager-769d9c7585-hpwgw\" (UID: \"29c9510f-29ed-482f-8b63-377479eff207\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.897170 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.897209 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv7bs\" (UniqueName: \"kubernetes.io/projected/4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2-kube-api-access-dv7bs\") pod \"watcher-operator-controller-manager-7cd4fb6f79-n728k\" (UID: \"4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.897250 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4z9m\" (UniqueName: \"kubernetes.io/projected/c72780f9-1a11-4ddc-bf65-3b1eb726be9e-kube-api-access-q4z9m\") pod \"test-operator-controller-manager-8464cf66df-s4mqg\" (UID: \"c72780f9-1a11-4ddc-bf65-3b1eb726be9e\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.909598 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29c9510f-29ed-482f-8b63-377479eff207-cert\") pod \"infra-operator-controller-manager-769d9c7585-hpwgw\" (UID: \"29c9510f-29ed-482f-8b63-377479eff207\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.930359 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.946259 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.948002 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.954983 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-67prf" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.961823 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5"] Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.962707 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4z9m\" (UniqueName: \"kubernetes.io/projected/c72780f9-1a11-4ddc-bf65-3b1eb726be9e-kube-api-access-q4z9m\") pod \"test-operator-controller-manager-8464cf66df-s4mqg\" (UID: \"c72780f9-1a11-4ddc-bf65-3b1eb726be9e\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.998409 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6n6d\" (UniqueName: \"kubernetes.io/projected/44c3b95a-a75b-4121-967a-28d483eb72a2-kube-api-access-x6n6d\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.998507 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.998552 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lb9c\" (UniqueName: \"kubernetes.io/projected/f09be5e8-3996-4aa0-956f-cd5b5cd0d08b-kube-api-access-4lb9c\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5\" (UID: \"f09be5e8-3996-4aa0-956f-cd5b5cd0d08b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" Nov 23 07:06:42 crc kubenswrapper[4906]: I1123 07:06:42.998579 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv7bs\" (UniqueName: \"kubernetes.io/projected/4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2-kube-api-access-dv7bs\") pod \"watcher-operator-controller-manager-7cd4fb6f79-n728k\" (UID: \"4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" Nov 23 07:06:42 crc kubenswrapper[4906]: E1123 07:06:42.999374 4906 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 23 07:06:42 crc kubenswrapper[4906]: E1123 07:06:42.999452 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert podName:44c3b95a-a75b-4121-967a-28d483eb72a2 nodeName:}" failed. No retries permitted until 2025-11-23 07:06:43.499415369 +0000 UTC m=+1019.012806672 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert") pod "openstack-operator-controller-manager-6cb9dc54f8-jbgnq" (UID: "44c3b95a-a75b-4121-967a-28d483eb72a2") : secret "webhook-server-cert" not found Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.010822 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv"] Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.049153 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.055238 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6n6d\" (UniqueName: \"kubernetes.io/projected/44c3b95a-a75b-4121-967a-28d483eb72a2-kube-api-access-x6n6d\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.058987 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.064183 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw"] Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.090594 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg"] Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.097455 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv7bs\" (UniqueName: \"kubernetes.io/projected/4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2-kube-api-access-dv7bs\") pod \"watcher-operator-controller-manager-7cd4fb6f79-n728k\" (UID: \"4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.100832 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d43ca54d-dde3-4178-bb66-1bf010c5dac8-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd4489m6x\" (UID: \"d43ca54d-dde3-4178-bb66-1bf010c5dac8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.100922 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lb9c\" (UniqueName: \"kubernetes.io/projected/f09be5e8-3996-4aa0-956f-cd5b5cd0d08b-kube-api-access-4lb9c\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5\" (UID: \"f09be5e8-3996-4aa0-956f-cd5b5cd0d08b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.104201 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.139307 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d43ca54d-dde3-4178-bb66-1bf010c5dac8-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd4489m6x\" (UID: \"d43ca54d-dde3-4178-bb66-1bf010c5dac8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.145222 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" event={"ID":"f7a10077-2d8a-4b42-8683-bfc0f336f7d1","Type":"ContainerStarted","Data":"3bb61ac3c9b07d1946cbcb585e82c29e65cd1050003cba217092819aecfcac21"} Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.145953 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lb9c\" (UniqueName: \"kubernetes.io/projected/f09be5e8-3996-4aa0-956f-cd5b5cd0d08b-kube-api-access-4lb9c\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5\" (UID: \"f09be5e8-3996-4aa0-956f-cd5b5cd0d08b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.286239 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.318662 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.323548 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6"] Nov 23 07:06:43 crc kubenswrapper[4906]: W1123 07:06:43.435743 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8480261_0abe_4b7d_b741_b76fe8126894.slice/crio-d768cc2ff5923dcebed8a271115d2d41208a4d6f0d8f14ea55b6861fa9926a9f WatchSource:0}: Error finding container d768cc2ff5923dcebed8a271115d2d41208a4d6f0d8f14ea55b6861fa9926a9f: Status 404 returned error can't find the container with id d768cc2ff5923dcebed8a271115d2d41208a4d6f0d8f14ea55b6861fa9926a9f Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.519938 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:43 crc kubenswrapper[4906]: E1123 07:06:43.520832 4906 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 23 07:06:43 crc kubenswrapper[4906]: E1123 07:06:43.520899 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert podName:44c3b95a-a75b-4121-967a-28d483eb72a2 nodeName:}" failed. No retries permitted until 2025-11-23 07:06:44.520878353 +0000 UTC m=+1020.034269656 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert") pod "openstack-operator-controller-manager-6cb9dc54f8-jbgnq" (UID: "44c3b95a-a75b-4121-967a-28d483eb72a2") : secret "webhook-server-cert" not found Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.755601 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n"] Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.761522 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm"] Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.962557 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv"] Nov 23 07:06:43 crc kubenswrapper[4906]: W1123 07:06:43.967596 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9bdc0d6_3a2e_4eda_8c73_870215a2018d.slice/crio-3731ef9e66b9ddace6e83a070431f36657bbace757af088e4310b3e15693cde4 WatchSource:0}: Error finding container 3731ef9e66b9ddace6e83a070431f36657bbace757af088e4310b3e15693cde4: Status 404 returned error can't find the container with id 3731ef9e66b9ddace6e83a070431f36657bbace757af088e4310b3e15693cde4 Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.970797 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f"] Nov 23 07:06:43 crc kubenswrapper[4906]: W1123 07:06:43.972926 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdd2bdf7_90ef_42f0_9395_b4889673d69c.slice/crio-0dc9de6286539517a7e092c8d51e943263e3544244059e3a3f31a824cbbf280c WatchSource:0}: Error finding container 0dc9de6286539517a7e092c8d51e943263e3544244059e3a3f31a824cbbf280c: Status 404 returned error can't find the container with id 0dc9de6286539517a7e092c8d51e943263e3544244059e3a3f31a824cbbf280c Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.976987 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt"] Nov 23 07:06:43 crc kubenswrapper[4906]: I1123 07:06:43.981702 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4"] Nov 23 07:06:43 crc kubenswrapper[4906]: W1123 07:06:43.994019 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45a30ceb_58e3_4e36_96a0_b64f35fa5d5e.slice/crio-c037c5a37d3b26b9110d390ffc72f7012c495e77a2efe82e38512e3e87204258 WatchSource:0}: Error finding container c037c5a37d3b26b9110d390ffc72f7012c495e77a2efe82e38512e3e87204258: Status 404 returned error can't find the container with id c037c5a37d3b26b9110d390ffc72f7012c495e77a2efe82e38512e3e87204258 Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.122796 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b"] Nov 23 07:06:44 crc kubenswrapper[4906]: W1123 07:06:44.128514 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd441d38_8fef_4ee4_8e92_f32682d524c3.slice/crio-0519f7997b8177d9c7c312a3b0565ae235413bd12ef684641541aa8aedbb2d53 WatchSource:0}: Error finding container 0519f7997b8177d9c7c312a3b0565ae235413bd12ef684641541aa8aedbb2d53: Status 404 returned error can't find the container with id 0519f7997b8177d9c7c312a3b0565ae235413bd12ef684641541aa8aedbb2d53 Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.135065 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm"] Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.144097 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f"] Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.149073 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j"] Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.163170 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr"] Nov 23 07:06:44 crc kubenswrapper[4906]: W1123 07:06:44.164292 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf050d900_ee68_4bdb_ad1f_3e64bf8db957.slice/crio-03de53e668d8b999f0aae01167925292eae984b6f16107d426136f70d0b063d1 WatchSource:0}: Error finding container 03de53e668d8b999f0aae01167925292eae984b6f16107d426136f70d0b063d1: Status 404 returned error can't find the container with id 03de53e668d8b999f0aae01167925292eae984b6f16107d426136f70d0b063d1 Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.168062 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" event={"ID":"45a30ceb-58e3-4e36-96a0-b64f35fa5d5e","Type":"ContainerStarted","Data":"c037c5a37d3b26b9110d390ffc72f7012c495e77a2efe82e38512e3e87204258"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.169257 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" event={"ID":"c4eaa84e-ff1f-4a9a-9ffe-56350431eb01","Type":"ContainerStarted","Data":"b88da248433335828646e3588a02e2d80587df7100ea3e75986ee6d2b25dadad"} Nov 23 07:06:44 crc kubenswrapper[4906]: W1123 07:06:44.170159 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod128c4cf7_46dc_4438_a13e_ebb14600b794.slice/crio-fcc8729de1c0348a2599ab25122681edff93c25713c924814bc819db8477be67 WatchSource:0}: Error finding container fcc8729de1c0348a2599ab25122681edff93c25713c924814bc819db8477be67: Status 404 returned error can't find the container with id fcc8729de1c0348a2599ab25122681edff93c25713c924814bc819db8477be67 Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.171547 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" event={"ID":"896605e2-1d27-4bda-a8bf-783a9fbd0294","Type":"ContainerStarted","Data":"f44f785a6df6cf226c6c7ff332e1aef10bda249d20c5b9f0323bf05d484279a1"} Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.173060 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l5qrj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-5bdf4f7f7f-gd2gm_openstack-operators(128c4cf7-46dc-4438-a13e-ebb14600b794): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.173511 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" event={"ID":"b9bdc0d6-3a2e-4eda-8c73-870215a2018d","Type":"ContainerStarted","Data":"3731ef9e66b9ddace6e83a070431f36657bbace757af088e4310b3e15693cde4"} Nov 23 07:06:44 crc kubenswrapper[4906]: W1123 07:06:44.175299 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ff2fa48_6adc_4bac_bf2d_a6a1b87f9ce2.slice/crio-455b15678f3b8a38fe5ad6d52629b590816ca740296a4e59b3410140192c3a5d WatchSource:0}: Error finding container 455b15678f3b8a38fe5ad6d52629b590816ca740296a4e59b3410140192c3a5d: Status 404 returned error can't find the container with id 455b15678f3b8a38fe5ad6d52629b590816ca740296a4e59b3410140192c3a5d Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.175810 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" event={"ID":"dd441d38-8fef-4ee4-8e92-f32682d524c3","Type":"ContainerStarted","Data":"0519f7997b8177d9c7c312a3b0565ae235413bd12ef684641541aa8aedbb2d53"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.176573 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm"] Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.177308 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dv7bs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7cd4fb6f79-n728k_openstack-operators(4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.180452 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" event={"ID":"f98f97de-580b-460d-9576-1b496313001f","Type":"ContainerStarted","Data":"eec512dac2059a2d45c73c6ae168d354a4b72118cc23df5d6b52ff22d9da15f7"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.183269 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k"] Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.185102 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" event={"ID":"e8480261-0abe-4b7d-b741-b76fe8126894","Type":"ContainerStarted","Data":"d768cc2ff5923dcebed8a271115d2d41208a4d6f0d8f14ea55b6861fa9926a9f"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.200754 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" event={"ID":"3bc07001-7205-48f6-a348-e62df3c098da","Type":"ContainerStarted","Data":"4da804ecc67f8b6e0a37b2617243d80c3b7c41168de7470e168f7e162835051f"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.203814 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" event={"ID":"150266f3-2d08-4b22-b00e-a04fa8419820","Type":"ContainerStarted","Data":"22c34fde22b07d06f6b0b2c03e73e810202b479ca0a75003520ca9015661fef4"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.211278 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" event={"ID":"cdd2bdf7-90ef-42f0-9395-b4889673d69c","Type":"ContainerStarted","Data":"0dc9de6286539517a7e092c8d51e943263e3544244059e3a3f31a824cbbf280c"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.215563 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" event={"ID":"bb4b51f3-5f0e-4a41-920c-7a179c1e0032","Type":"ContainerStarted","Data":"c8c23e0b7850c92facdf9b0d16a337664cd243448bd3c9c5a13fc20fb0b5c5d2"} Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.318580 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7"] Nov 23 07:06:44 crc kubenswrapper[4906]: W1123 07:06:44.325982 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29c9510f_29ed_482f_8b63_377479eff207.slice/crio-26fce1f05ff20fe357a2c890bd6def115136df79c0688fe1f612caca2ae8aa7c WatchSource:0}: Error finding container 26fce1f05ff20fe357a2c890bd6def115136df79c0688fe1f612caca2ae8aa7c: Status 404 returned error can't find the container with id 26fce1f05ff20fe357a2c890bd6def115136df79c0688fe1f612caca2ae8aa7c Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.336164 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bm74m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-769d9c7585-hpwgw_openstack-operators(29c9510f-29ed-482f-8b63-377479eff207): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.337108 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q4z9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-8464cf66df-s4mqg_openstack-operators(c72780f9-1a11-4ddc-bf65-3b1eb726be9e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.337539 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw"] Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.349657 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg"] Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.356996 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5"] Nov 23 07:06:44 crc kubenswrapper[4906]: W1123 07:06:44.360186 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd43ca54d_dde3_4178_bb66_1bf010c5dac8.slice/crio-d0677796f54d7fb857a390616e9e0ec3aba5476ed0e9cf8e15865436113cdd76 WatchSource:0}: Error finding container d0677796f54d7fb857a390616e9e0ec3aba5476ed0e9cf8e15865436113cdd76: Status 404 returned error can't find the container with id d0677796f54d7fb857a390616e9e0ec3aba5476ed0e9cf8e15865436113cdd76 Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.363332 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x"] Nov 23 07:06:44 crc kubenswrapper[4906]: W1123 07:06:44.381664 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod962c14dd_6e7a_4927_ba5b_84384bde4312.slice/crio-2f7f781f475a35b25e0721fd3d11817c76a22761002505f440f6d9856c455ab4 WatchSource:0}: Error finding container 2f7f781f475a35b25e0721fd3d11817c76a22761002505f440f6d9856c455ab4: Status 404 returned error can't find the container with id 2f7f781f475a35b25e0721fd3d11817c76a22761002505f440f6d9856c455ab4 Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.401144 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4lb9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5_openstack-operators(f09be5e8-3996-4aa0-956f-cd5b5cd0d08b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.401803 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ljzwz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-86d796d84d-kqzb7_openstack-operators(962c14dd-6e7a-4927-ba5b-84384bde4312): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.402942 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" podUID="f09be5e8-3996-4aa0-956f-cd5b5cd0d08b" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.409905 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" podUID="4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.446831 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" podUID="128c4cf7-46dc-4438-a13e-ebb14600b794" Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.551023 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.559100 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/44c3b95a-a75b-4121-967a-28d483eb72a2-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-jbgnq\" (UID: \"44c3b95a-a75b-4121-967a-28d483eb72a2\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.574231 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" podUID="c72780f9-1a11-4ddc-bf65-3b1eb726be9e" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.586717 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" podUID="29c9510f-29ed-482f-8b63-377479eff207" Nov 23 07:06:44 crc kubenswrapper[4906]: E1123 07:06:44.600126 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" podUID="962c14dd-6e7a-4927-ba5b-84384bde4312" Nov 23 07:06:44 crc kubenswrapper[4906]: I1123 07:06:44.685302 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.249704 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" event={"ID":"3247edc2-e84e-4c6b-9b60-cbfb7a7ec442","Type":"ContainerStarted","Data":"f9d7ea7e7e838bfdcc32eee5ff66b8bea7f76f59fbe37f63f02d50fba5ebb95e"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.253208 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" event={"ID":"29c9510f-29ed-482f-8b63-377479eff207","Type":"ContainerStarted","Data":"51819c8fc68280470d7f7f503f3601ed2eb69e5e663de23679039f34a0dc4332"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.253251 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" event={"ID":"29c9510f-29ed-482f-8b63-377479eff207","Type":"ContainerStarted","Data":"26fce1f05ff20fe357a2c890bd6def115136df79c0688fe1f612caca2ae8aa7c"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.259813 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" event={"ID":"4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2","Type":"ContainerStarted","Data":"101038f4746237a00bdfc15457d6df2d06f0a5c91a255911668af51ae609b7a6"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.259877 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" event={"ID":"4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2","Type":"ContainerStarted","Data":"455b15678f3b8a38fe5ad6d52629b590816ca740296a4e59b3410140192c3a5d"} Nov 23 07:06:45 crc kubenswrapper[4906]: E1123 07:06:45.266278 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" podUID="29c9510f-29ed-482f-8b63-377479eff207" Nov 23 07:06:45 crc kubenswrapper[4906]: E1123 07:06:45.266363 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" podUID="4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2" Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.291745 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" event={"ID":"c72780f9-1a11-4ddc-bf65-3b1eb726be9e","Type":"ContainerStarted","Data":"3819565efb576e8040c062c0d68f835ec62b9c3a25d8cd7571d4eeb7c51ccf5d"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.291800 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" event={"ID":"c72780f9-1a11-4ddc-bf65-3b1eb726be9e","Type":"ContainerStarted","Data":"94d82aa0ea8a285d9a7b9fe2d6c4c8008378e03d107f63b3faa5568ff264571d"} Nov 23 07:06:45 crc kubenswrapper[4906]: E1123 07:06:45.297713 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" podUID="c72780f9-1a11-4ddc-bf65-3b1eb726be9e" Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.299064 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" event={"ID":"f050d900-ee68-4bdb-ad1f-3e64bf8db957","Type":"ContainerStarted","Data":"03de53e668d8b999f0aae01167925292eae984b6f16107d426136f70d0b063d1"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.312433 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" event={"ID":"2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239","Type":"ContainerStarted","Data":"dd58ffaa6a64361eaa8b51ac4a532595c4f192b5c19f8694be823708875c799f"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.322275 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" event={"ID":"d43ca54d-dde3-4178-bb66-1bf010c5dac8","Type":"ContainerStarted","Data":"d0677796f54d7fb857a390616e9e0ec3aba5476ed0e9cf8e15865436113cdd76"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.329661 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" event={"ID":"962c14dd-6e7a-4927-ba5b-84384bde4312","Type":"ContainerStarted","Data":"f3693717ac4b102eed2aeb7fee98d9b5712df44ec9a3595321cbecabe7a9e6fc"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.329717 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" event={"ID":"962c14dd-6e7a-4927-ba5b-84384bde4312","Type":"ContainerStarted","Data":"2f7f781f475a35b25e0721fd3d11817c76a22761002505f440f6d9856c455ab4"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.331723 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq"] Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.332072 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" event={"ID":"128c4cf7-46dc-4438-a13e-ebb14600b794","Type":"ContainerStarted","Data":"a331af68f92638023e364994c12c17b5705f460f980d8ba7d926852a01fe0cc2"} Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.332130 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" event={"ID":"128c4cf7-46dc-4438-a13e-ebb14600b794","Type":"ContainerStarted","Data":"fcc8729de1c0348a2599ab25122681edff93c25713c924814bc819db8477be67"} Nov 23 07:06:45 crc kubenswrapper[4906]: E1123 07:06:45.334240 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" podUID="128c4cf7-46dc-4438-a13e-ebb14600b794" Nov 23 07:06:45 crc kubenswrapper[4906]: I1123 07:06:45.335930 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" event={"ID":"f09be5e8-3996-4aa0-956f-cd5b5cd0d08b","Type":"ContainerStarted","Data":"1364e9321da58c0e68ed9f0ab3075519886083e3e947acf69e9efc9884b86a63"} Nov 23 07:06:45 crc kubenswrapper[4906]: E1123 07:06:45.336107 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" podUID="962c14dd-6e7a-4927-ba5b-84384bde4312" Nov 23 07:06:45 crc kubenswrapper[4906]: E1123 07:06:45.348828 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" podUID="f09be5e8-3996-4aa0-956f-cd5b5cd0d08b" Nov 23 07:06:46 crc kubenswrapper[4906]: E1123 07:06:46.345143 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" podUID="4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2" Nov 23 07:06:46 crc kubenswrapper[4906]: E1123 07:06:46.345957 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" podUID="c72780f9-1a11-4ddc-bf65-3b1eb726be9e" Nov 23 07:06:46 crc kubenswrapper[4906]: E1123 07:06:46.346492 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" podUID="962c14dd-6e7a-4927-ba5b-84384bde4312" Nov 23 07:06:46 crc kubenswrapper[4906]: E1123 07:06:46.346587 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" podUID="29c9510f-29ed-482f-8b63-377479eff207" Nov 23 07:06:46 crc kubenswrapper[4906]: E1123 07:06:46.346829 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" podUID="128c4cf7-46dc-4438-a13e-ebb14600b794" Nov 23 07:06:46 crc kubenswrapper[4906]: E1123 07:06:46.346921 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" podUID="f09be5e8-3996-4aa0-956f-cd5b5cd0d08b" Nov 23 07:06:50 crc kubenswrapper[4906]: W1123 07:06:50.343335 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44c3b95a_a75b_4121_967a_28d483eb72a2.slice/crio-e80d9a06835963ea4aead66a23423653b169f4b6af0c8fe7390d98f0d5a2177b WatchSource:0}: Error finding container e80d9a06835963ea4aead66a23423653b169f4b6af0c8fe7390d98f0d5a2177b: Status 404 returned error can't find the container with id e80d9a06835963ea4aead66a23423653b169f4b6af0c8fe7390d98f0d5a2177b Nov 23 07:06:50 crc kubenswrapper[4906]: I1123 07:06:50.373486 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" event={"ID":"44c3b95a-a75b-4121-967a-28d483eb72a2","Type":"ContainerStarted","Data":"e80d9a06835963ea4aead66a23423653b169f4b6af0c8fe7390d98f0d5a2177b"} Nov 23 07:06:50 crc kubenswrapper[4906]: I1123 07:06:50.946082 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:06:50 crc kubenswrapper[4906]: I1123 07:06:50.946165 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.477358 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" event={"ID":"e8480261-0abe-4b7d-b741-b76fe8126894","Type":"ContainerStarted","Data":"71f292355cfbe8c57870c8aaf56ab4363a730e18c44cb3effeefc67f1f5941ef"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.491500 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" event={"ID":"bb4b51f3-5f0e-4a41-920c-7a179c1e0032","Type":"ContainerStarted","Data":"5c6f44c73c6ab4642152bb03df4812a2cd25ae0b88eb81b363da57f97125ae46"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.493109 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" event={"ID":"c4eaa84e-ff1f-4a9a-9ffe-56350431eb01","Type":"ContainerStarted","Data":"03b155afbf842545affb654b65cee6e5a9acab8efb3e687990d31d907ce024ca"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.494399 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" event={"ID":"44c3b95a-a75b-4121-967a-28d483eb72a2","Type":"ContainerStarted","Data":"cb6aceaa5189944adb498a76f3f403bc244d9e77b2d35bf3dfaf66ceaf87d5d7"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.494452 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" event={"ID":"44c3b95a-a75b-4121-967a-28d483eb72a2","Type":"ContainerStarted","Data":"42d9c67381bb38d6b1e77836ee2e44a81f35e5b8cb9907756c4ee022e44ce4b2"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.495027 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.505317 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" event={"ID":"f7a10077-2d8a-4b42-8683-bfc0f336f7d1","Type":"ContainerStarted","Data":"017d3250e62aebe272589ac86dd8dcc951b51720ccc5340a565601538992d5c7"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.519855 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" event={"ID":"2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239","Type":"ContainerStarted","Data":"022d0a837355f2947286c6c3d7a17e29de661ace341c5e7ef9c35eb7e8bcaf6f"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.521896 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" event={"ID":"d43ca54d-dde3-4178-bb66-1bf010c5dac8","Type":"ContainerStarted","Data":"48a20d763721ea154bc1a3109712035dc5c983a1d09d8a0dd9bc6e9f5176cfb4"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.530554 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" event={"ID":"45a30ceb-58e3-4e36-96a0-b64f35fa5d5e","Type":"ContainerStarted","Data":"3801c9e4369443e2cd9bbb7af82c59fc6b12156b49f1ef7d94bd88cb8fdf99ef"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.546088 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" event={"ID":"f98f97de-580b-460d-9576-1b496313001f","Type":"ContainerStarted","Data":"f936b7cf5e227209e34c2706c38f5cb8700ec6ccd49156d00fbc912cec96d554"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.565257 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" event={"ID":"3247edc2-e84e-4c6b-9b60-cbfb7a7ec442","Type":"ContainerStarted","Data":"dcc7f3fb50edf06f5d8bcbe54ac727624138536d4dd58a0fc7e818e00cfaff45"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.586766 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" podStartSLOduration=13.58674992 podStartE2EDuration="13.58674992s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:06:55.564198719 +0000 UTC m=+1031.077590022" watchObservedRunningTime="2025-11-23 07:06:55.58674992 +0000 UTC m=+1031.100141223" Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.598414 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" event={"ID":"b9bdc0d6-3a2e-4eda-8c73-870215a2018d","Type":"ContainerStarted","Data":"8fb7ba689c5b251a31c1ff9c81f6532d881706809bdd5b54cd8aa9707ad9569d"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.611313 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" event={"ID":"dd441d38-8fef-4ee4-8e92-f32682d524c3","Type":"ContainerStarted","Data":"87e9a3f57d58e51cbb6027586bbc42e02025dc7e8bebb17f635c42aa57dd48b2"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.629998 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" event={"ID":"896605e2-1d27-4bda-a8bf-783a9fbd0294","Type":"ContainerStarted","Data":"ed804405208d1147ff00f3ed8d04653948804f1ef2f2550e01810c21ede879b7"} Nov 23 07:06:55 crc kubenswrapper[4906]: I1123 07:06:55.643083 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" event={"ID":"f050d900-ee68-4bdb-ad1f-3e64bf8db957","Type":"ContainerStarted","Data":"07e14f6d135ace7e13a76511522c1ac185b253fdfedbd58f2c00c44eaf0791ab"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.655395 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" event={"ID":"cdd2bdf7-90ef-42f0-9395-b4889673d69c","Type":"ContainerStarted","Data":"038a1ac4f7b1a62d877955e2898f831fb05de9b7452177b7c40f566af135c416"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.655859 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" event={"ID":"cdd2bdf7-90ef-42f0-9395-b4889673d69c","Type":"ContainerStarted","Data":"7452983e5389196f20cac875464768af0c4e43cf78ceb8ba1bf03120e65f3f73"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.655885 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.657710 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" event={"ID":"e8480261-0abe-4b7d-b741-b76fe8126894","Type":"ContainerStarted","Data":"eead3c3d28cf6e95e0cdf51b8efa8e7401264887373ae68c0bdf95ffcd25c563"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.658246 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.663179 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" event={"ID":"2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239","Type":"ContainerStarted","Data":"4c982b05654073a37d237d104af9edf75b47ba2d84ec8ced57eb196b530b0164"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.663303 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.665771 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" event={"ID":"3247edc2-e84e-4c6b-9b60-cbfb7a7ec442","Type":"ContainerStarted","Data":"e01e1aa1c000c98f15d3e15996169236854dfd9f6c0adb233de16ec483822b2a"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.665921 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.669186 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" event={"ID":"896605e2-1d27-4bda-a8bf-783a9fbd0294","Type":"ContainerStarted","Data":"19c19c9cbbbe66adef16731d2be27fb5a9ae47d6dab247343b6f56afb65cafad"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.669325 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.671455 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" event={"ID":"b9bdc0d6-3a2e-4eda-8c73-870215a2018d","Type":"ContainerStarted","Data":"3871a7f34ec00b8055118a9f2c8112049190ad3a7c2028544ed035960068e41b"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.671889 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.680432 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" podStartSLOduration=4.119824557 podStartE2EDuration="14.68041913s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.976203504 +0000 UTC m=+1019.489594807" lastFinishedPulling="2025-11-23 07:06:54.536798077 +0000 UTC m=+1030.050189380" observedRunningTime="2025-11-23 07:06:56.672496502 +0000 UTC m=+1032.185887805" watchObservedRunningTime="2025-11-23 07:06:56.68041913 +0000 UTC m=+1032.193810433" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.686888 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.693264 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" event={"ID":"45a30ceb-58e3-4e36-96a0-b64f35fa5d5e","Type":"ContainerStarted","Data":"98490f7fc98bb0c6ddecf3a7fc9a5f4754534527ba5495bd22d54b7ab22c2240"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.693800 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.700586 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" podStartSLOduration=5.161958691 podStartE2EDuration="15.700572168s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.996936377 +0000 UTC m=+1019.510327680" lastFinishedPulling="2025-11-23 07:06:54.535549824 +0000 UTC m=+1030.048941157" observedRunningTime="2025-11-23 07:06:56.694421767 +0000 UTC m=+1032.207813070" watchObservedRunningTime="2025-11-23 07:06:56.700572168 +0000 UTC m=+1032.213963471" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.710521 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" event={"ID":"c4eaa84e-ff1f-4a9a-9ffe-56350431eb01","Type":"ContainerStarted","Data":"9a2f20ab7fc2018ecc9bf1f8c0766ec2503425bbd3b8bb277cb103aac974ac0d"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.711089 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.725467 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" event={"ID":"f7a10077-2d8a-4b42-8683-bfc0f336f7d1","Type":"ContainerStarted","Data":"9ae37252c1fe5257bbe82de358204bcba516db3fd2912ad9539345ce66a123cc"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.725547 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.727325 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" podStartSLOduration=4.162114326 podStartE2EDuration="14.72730173s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.971319706 +0000 UTC m=+1019.484711009" lastFinishedPulling="2025-11-23 07:06:54.53650711 +0000 UTC m=+1030.049898413" observedRunningTime="2025-11-23 07:06:56.720162292 +0000 UTC m=+1032.233553615" watchObservedRunningTime="2025-11-23 07:06:56.72730173 +0000 UTC m=+1032.240693033" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.736759 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" event={"ID":"150266f3-2d08-4b22-b00e-a04fa8419820","Type":"ContainerStarted","Data":"b0bb42d8b5efffd22c7bc64ac57fdbca7302dcbbef1f331c2b224a9ed58bbdbf"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.736838 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" event={"ID":"150266f3-2d08-4b22-b00e-a04fa8419820","Type":"ContainerStarted","Data":"cd9b0a3e5d27ee1fc9c25cfd742c2f1d8b14be4d63db61192d0f11384d41a98c"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.737346 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.745811 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" podStartSLOduration=4.374783613 podStartE2EDuration="14.745791384s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.164544133 +0000 UTC m=+1019.677935436" lastFinishedPulling="2025-11-23 07:06:54.535551904 +0000 UTC m=+1030.048943207" observedRunningTime="2025-11-23 07:06:56.744029347 +0000 UTC m=+1032.257420650" watchObservedRunningTime="2025-11-23 07:06:56.745791384 +0000 UTC m=+1032.259182687" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.746874 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" event={"ID":"d43ca54d-dde3-4178-bb66-1bf010c5dac8","Type":"ContainerStarted","Data":"acbed0ead568049e59e1689dd5e2b408668548a53ca8030824001e4773ab6ffa"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.747925 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.756866 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.767130 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" event={"ID":"f050d900-ee68-4bdb-ad1f-3e64bf8db957","Type":"ContainerStarted","Data":"e54c1cc1e3f63cde076c6adb7504483386ec8c125e1e6fba914adef7550f7403"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.767343 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.769911 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" event={"ID":"3bc07001-7205-48f6-a348-e62df3c098da","Type":"ContainerStarted","Data":"f69922a35fca68f4946922ee0e245c96319181c44bb4ebbebe9194c5807a026a"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.769996 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.770020 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" event={"ID":"3bc07001-7205-48f6-a348-e62df3c098da","Type":"ContainerStarted","Data":"a46a24946b3a6b64b779e83949e9185317468071b03621c028542ab7e7943dd2"} Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.778768 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" podStartSLOduration=4.69822515 podStartE2EDuration="15.778747288s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.456238288 +0000 UTC m=+1018.969629591" lastFinishedPulling="2025-11-23 07:06:54.536760386 +0000 UTC m=+1030.050151729" observedRunningTime="2025-11-23 07:06:56.774467746 +0000 UTC m=+1032.287859049" watchObservedRunningTime="2025-11-23 07:06:56.778747288 +0000 UTC m=+1032.292138591" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.806246 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" podStartSLOduration=4.43377135 podStartE2EDuration="14.806219839s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.164401649 +0000 UTC m=+1019.677792952" lastFinishedPulling="2025-11-23 07:06:54.536850138 +0000 UTC m=+1030.050241441" observedRunningTime="2025-11-23 07:06:56.803401055 +0000 UTC m=+1032.316792358" watchObservedRunningTime="2025-11-23 07:06:56.806219839 +0000 UTC m=+1032.319611142" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.828679 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" podStartSLOduration=4.532112805 podStartE2EDuration="15.828645997s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.240245515 +0000 UTC m=+1018.753636818" lastFinishedPulling="2025-11-23 07:06:54.536778697 +0000 UTC m=+1030.050170010" observedRunningTime="2025-11-23 07:06:56.827988089 +0000 UTC m=+1032.341379402" watchObservedRunningTime="2025-11-23 07:06:56.828645997 +0000 UTC m=+1032.342037300" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.859968 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" podStartSLOduration=4.709268754 podStartE2EDuration="14.859949467s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.376792669 +0000 UTC m=+1019.890183972" lastFinishedPulling="2025-11-23 07:06:54.527473382 +0000 UTC m=+1030.040864685" observedRunningTime="2025-11-23 07:06:56.855641575 +0000 UTC m=+1032.369032878" watchObservedRunningTime="2025-11-23 07:06:56.859949467 +0000 UTC m=+1032.373340770" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.885525 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" podStartSLOduration=4.490783732 podStartE2EDuration="15.885500388s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.145345136 +0000 UTC m=+1018.658736439" lastFinishedPulling="2025-11-23 07:06:54.540061762 +0000 UTC m=+1030.053453095" observedRunningTime="2025-11-23 07:06:56.88101984 +0000 UTC m=+1032.394411143" watchObservedRunningTime="2025-11-23 07:06:56.885500388 +0000 UTC m=+1032.398891691" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.905766 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" podStartSLOduration=4.514197307 podStartE2EDuration="14.905738058s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.162769546 +0000 UTC m=+1019.676160849" lastFinishedPulling="2025-11-23 07:06:54.554310297 +0000 UTC m=+1030.067701600" observedRunningTime="2025-11-23 07:06:56.899711601 +0000 UTC m=+1032.413102904" watchObservedRunningTime="2025-11-23 07:06:56.905738058 +0000 UTC m=+1032.419129381" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.936209 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" podStartSLOduration=4.533106355 podStartE2EDuration="14.936180867s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.13431423 +0000 UTC m=+1019.647705533" lastFinishedPulling="2025-11-23 07:06:54.537388732 +0000 UTC m=+1030.050780045" observedRunningTime="2025-11-23 07:06:56.934827641 +0000 UTC m=+1032.448218944" watchObservedRunningTime="2025-11-23 07:06:56.936180867 +0000 UTC m=+1032.449572170" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.952346 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" podStartSLOduration=4.413149229 podStartE2EDuration="14.95231854s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.99739134 +0000 UTC m=+1019.510782643" lastFinishedPulling="2025-11-23 07:06:54.536560631 +0000 UTC m=+1030.049951954" observedRunningTime="2025-11-23 07:06:56.951490569 +0000 UTC m=+1032.464881872" watchObservedRunningTime="2025-11-23 07:06:56.95231854 +0000 UTC m=+1032.465709843" Nov 23 07:06:56 crc kubenswrapper[4906]: I1123 07:06:56.976286 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" podStartSLOduration=4.607733031 podStartE2EDuration="14.976261847s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.168309542 +0000 UTC m=+1019.681700845" lastFinishedPulling="2025-11-23 07:06:54.536838318 +0000 UTC m=+1030.050229661" observedRunningTime="2025-11-23 07:06:56.973823874 +0000 UTC m=+1032.487215177" watchObservedRunningTime="2025-11-23 07:06:56.976261847 +0000 UTC m=+1032.489653150" Nov 23 07:06:57 crc kubenswrapper[4906]: I1123 07:06:57.002645 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" podStartSLOduration=4.800098202 podStartE2EDuration="16.002622119s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.336695224 +0000 UTC m=+1018.850086527" lastFinishedPulling="2025-11-23 07:06:54.539219141 +0000 UTC m=+1030.052610444" observedRunningTime="2025-11-23 07:06:56.998194633 +0000 UTC m=+1032.511585936" watchObservedRunningTime="2025-11-23 07:06:57.002622119 +0000 UTC m=+1032.516013422" Nov 23 07:06:57 crc kubenswrapper[4906]: I1123 07:06:57.032438 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" podStartSLOduration=4.261626705 podStartE2EDuration="15.03241698s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.765057557 +0000 UTC m=+1019.278448860" lastFinishedPulling="2025-11-23 07:06:54.535847832 +0000 UTC m=+1030.049239135" observedRunningTime="2025-11-23 07:06:57.024126703 +0000 UTC m=+1032.537518006" watchObservedRunningTime="2025-11-23 07:06:57.03241698 +0000 UTC m=+1032.545808283" Nov 23 07:06:57 crc kubenswrapper[4906]: I1123 07:06:57.781595 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" event={"ID":"dd441d38-8fef-4ee4-8e92-f32682d524c3","Type":"ContainerStarted","Data":"ca58d3efdb75bf29971133e8a7378a5cf72b08abd38bdd7b386d4e637765c085"} Nov 23 07:06:57 crc kubenswrapper[4906]: I1123 07:06:57.785311 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" event={"ID":"f98f97de-580b-460d-9576-1b496313001f","Type":"ContainerStarted","Data":"624fd9f970eab28bbaf194820fb2649a1c73c9e708f3582b324e3d1db1cd4480"} Nov 23 07:06:57 crc kubenswrapper[4906]: I1123 07:06:57.788855 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" event={"ID":"bb4b51f3-5f0e-4a41-920c-7a179c1e0032","Type":"ContainerStarted","Data":"c1e2ec08bc71aefea46ff6b344860f84d490532d2ce45c64a7815d5db6723ccb"} Nov 23 07:06:57 crc kubenswrapper[4906]: I1123 07:06:57.790647 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" Nov 23 07:06:57 crc kubenswrapper[4906]: I1123 07:06:57.816266 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" podStartSLOduration=6.047828422 podStartE2EDuration="16.816245725s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:43.768812585 +0000 UTC m=+1019.282203888" lastFinishedPulling="2025-11-23 07:06:54.537229888 +0000 UTC m=+1030.050621191" observedRunningTime="2025-11-23 07:06:57.810003811 +0000 UTC m=+1033.323395114" watchObservedRunningTime="2025-11-23 07:06:57.816245725 +0000 UTC m=+1033.329637028" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.257959 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-fkvbv" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.277370 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-59wgw" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.301238 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.374918 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-q6mmt" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.463883 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-jngn6" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.481034 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-vqg2n" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.549653 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-8tg5j" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.562016 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-xvpjm" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.587969 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-xm76f" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.653563 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-c9xlv" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.665310 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-7d26f" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.707287 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-q98m4" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.852550 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-ndc2b" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.882614 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-wtlrm" Nov 23 07:07:02 crc kubenswrapper[4906]: I1123 07:07:02.952172 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-f4jtr" Nov 23 07:07:03 crc kubenswrapper[4906]: I1123 07:07:03.334451 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd4489m6x" Nov 23 07:07:04 crc kubenswrapper[4906]: I1123 07:07:04.693132 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-jbgnq" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.926777 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" event={"ID":"29c9510f-29ed-482f-8b63-377479eff207","Type":"ContainerStarted","Data":"14e9b06f5d4f081fa9a1ef6041b82876c04142ae26850d8870375c3cdeaa93db"} Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.928888 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.931529 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" event={"ID":"4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2","Type":"ContainerStarted","Data":"29110267776d0671041701b05a70a692f3e24d32f854d1b6eab6c340a3151d26"} Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.932029 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.935962 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" event={"ID":"c72780f9-1a11-4ddc-bf65-3b1eb726be9e","Type":"ContainerStarted","Data":"e2b51d663c703b446a0c2f2541f78b621c0e975365de13eacc1966346246f903"} Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.936207 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.938248 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" event={"ID":"962c14dd-6e7a-4927-ba5b-84384bde4312","Type":"ContainerStarted","Data":"f2810d39d0c1894abc479c1cd12960a9f41b0f77ac7acfc0bc3f020c79c1a380"} Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.938473 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.940610 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" event={"ID":"128c4cf7-46dc-4438-a13e-ebb14600b794","Type":"ContainerStarted","Data":"bb80fe93b3317fcf3eaa078ba630d0a6b109806dd07d611e5f0460a4761c3b0c"} Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.940838 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.942690 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" event={"ID":"f09be5e8-3996-4aa0-956f-cd5b5cd0d08b","Type":"ContainerStarted","Data":"5bf0c67ea8d9f4131d02bb3417aa650e6ad4771de62549a315403cf39680d544"} Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.957146 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" podStartSLOduration=4.646161637 podStartE2EDuration="29.957113452s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.33601923 +0000 UTC m=+1019.849410533" lastFinishedPulling="2025-11-23 07:07:09.646971045 +0000 UTC m=+1045.160362348" observedRunningTime="2025-11-23 07:07:10.951391642 +0000 UTC m=+1046.464782945" watchObservedRunningTime="2025-11-23 07:07:10.957113452 +0000 UTC m=+1046.470504795" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.969910 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" podStartSLOduration=4.287466463 podStartE2EDuration="28.969876077s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.401559338 +0000 UTC m=+1019.914950641" lastFinishedPulling="2025-11-23 07:07:09.083968932 +0000 UTC m=+1044.597360255" observedRunningTime="2025-11-23 07:07:10.966867838 +0000 UTC m=+1046.480259141" watchObservedRunningTime="2025-11-23 07:07:10.969876077 +0000 UTC m=+1046.483267380" Nov 23 07:07:10 crc kubenswrapper[4906]: I1123 07:07:10.988774 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5" podStartSLOduration=3.808200175 podStartE2EDuration="28.988746271s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.400965413 +0000 UTC m=+1019.914356716" lastFinishedPulling="2025-11-23 07:07:09.581511509 +0000 UTC m=+1045.094902812" observedRunningTime="2025-11-23 07:07:10.985202399 +0000 UTC m=+1046.498593782" watchObservedRunningTime="2025-11-23 07:07:10.988746271 +0000 UTC m=+1046.502137574" Nov 23 07:07:11 crc kubenswrapper[4906]: I1123 07:07:11.008246 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" podStartSLOduration=3.504881051 podStartE2EDuration="29.008223612s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.172869081 +0000 UTC m=+1019.686260384" lastFinishedPulling="2025-11-23 07:07:09.676211642 +0000 UTC m=+1045.189602945" observedRunningTime="2025-11-23 07:07:11.007458572 +0000 UTC m=+1046.520849915" watchObservedRunningTime="2025-11-23 07:07:11.008223612 +0000 UTC m=+1046.521614915" Nov 23 07:07:11 crc kubenswrapper[4906]: I1123 07:07:11.031714 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" podStartSLOduration=3.526558629 podStartE2EDuration="29.031693527s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.177148943 +0000 UTC m=+1019.690540236" lastFinishedPulling="2025-11-23 07:07:09.682283831 +0000 UTC m=+1045.195675134" observedRunningTime="2025-11-23 07:07:11.029385257 +0000 UTC m=+1046.542776560" watchObservedRunningTime="2025-11-23 07:07:11.031693527 +0000 UTC m=+1046.545084831" Nov 23 07:07:11 crc kubenswrapper[4906]: I1123 07:07:11.053111 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" podStartSLOduration=3.743063076 podStartE2EDuration="29.053069828s" podCreationTimestamp="2025-11-23 07:06:42 +0000 UTC" firstStartedPulling="2025-11-23 07:06:44.337020925 +0000 UTC m=+1019.850412218" lastFinishedPulling="2025-11-23 07:07:09.647027657 +0000 UTC m=+1045.160418970" observedRunningTime="2025-11-23 07:07:11.047607384 +0000 UTC m=+1046.560998687" watchObservedRunningTime="2025-11-23 07:07:11.053069828 +0000 UTC m=+1046.566461171" Nov 23 07:07:20 crc kubenswrapper[4906]: I1123 07:07:20.946326 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:07:20 crc kubenswrapper[4906]: I1123 07:07:20.947278 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:07:20 crc kubenswrapper[4906]: I1123 07:07:20.947367 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:07:20 crc kubenswrapper[4906]: I1123 07:07:20.948729 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c632a9b1605581f702cb4b0b6cbc68235936754af5501ce83c90241ccf464cb9"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:07:20 crc kubenswrapper[4906]: I1123 07:07:20.948853 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://c632a9b1605581f702cb4b0b6cbc68235936754af5501ce83c90241ccf464cb9" gracePeriod=600 Nov 23 07:07:21 crc kubenswrapper[4906]: E1123 07:07:21.081620 4906 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec5c6569_64d4_4591_bbac_b70ebcca836f.slice/crio-conmon-c632a9b1605581f702cb4b0b6cbc68235936754af5501ce83c90241ccf464cb9.scope\": RecentStats: unable to find data in memory cache]" Nov 23 07:07:22 crc kubenswrapper[4906]: I1123 07:07:22.062267 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="c632a9b1605581f702cb4b0b6cbc68235936754af5501ce83c90241ccf464cb9" exitCode=0 Nov 23 07:07:22 crc kubenswrapper[4906]: I1123 07:07:22.062390 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"c632a9b1605581f702cb4b0b6cbc68235936754af5501ce83c90241ccf464cb9"} Nov 23 07:07:22 crc kubenswrapper[4906]: I1123 07:07:22.062921 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"cde0743cc2b7d9547c3784b89c951ad5af5f1a5ec9c54292ba0b98af998c4c47"} Nov 23 07:07:22 crc kubenswrapper[4906]: I1123 07:07:22.062963 4906 scope.go:117] "RemoveContainer" containerID="5e4be889a6576c0c1dc1c3392a1a758863587cc0404504906531a0c6425bf526" Nov 23 07:07:22 crc kubenswrapper[4906]: I1123 07:07:22.675158 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-kqzb7" Nov 23 07:07:22 crc kubenswrapper[4906]: I1123 07:07:22.709845 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-gd2gm" Nov 23 07:07:23 crc kubenswrapper[4906]: I1123 07:07:23.052803 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8464cf66df-s4mqg" Nov 23 07:07:23 crc kubenswrapper[4906]: I1123 07:07:23.069065 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-hpwgw" Nov 23 07:07:23 crc kubenswrapper[4906]: I1123 07:07:23.114782 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-n728k" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.327725 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-zhtcl"] Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.329822 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.331699 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.331928 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.332353 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.339077 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-h5gm8" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.355242 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-zhtcl"] Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.396952 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6584b49599-88dtq"] Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.401242 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.403025 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.410785 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-88dtq"] Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.505165 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-259sk\" (UniqueName: \"kubernetes.io/projected/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-kube-api-access-259sk\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.505235 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppccf\" (UniqueName: \"kubernetes.io/projected/c5e77d8e-c265-47e5-ac76-985caeb38679-kube-api-access-ppccf\") pod \"dnsmasq-dns-7bdd77c89-zhtcl\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.505267 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5e77d8e-c265-47e5-ac76-985caeb38679-config\") pod \"dnsmasq-dns-7bdd77c89-zhtcl\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.505337 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-config\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.505385 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-dns-svc\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.607496 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-259sk\" (UniqueName: \"kubernetes.io/projected/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-kube-api-access-259sk\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.607553 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppccf\" (UniqueName: \"kubernetes.io/projected/c5e77d8e-c265-47e5-ac76-985caeb38679-kube-api-access-ppccf\") pod \"dnsmasq-dns-7bdd77c89-zhtcl\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.607588 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5e77d8e-c265-47e5-ac76-985caeb38679-config\") pod \"dnsmasq-dns-7bdd77c89-zhtcl\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.607647 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-config\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.607694 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-dns-svc\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.608847 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-dns-svc\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.608864 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5e77d8e-c265-47e5-ac76-985caeb38679-config\") pod \"dnsmasq-dns-7bdd77c89-zhtcl\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.608985 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-config\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.631016 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppccf\" (UniqueName: \"kubernetes.io/projected/c5e77d8e-c265-47e5-ac76-985caeb38679-kube-api-access-ppccf\") pod \"dnsmasq-dns-7bdd77c89-zhtcl\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.632849 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-259sk\" (UniqueName: \"kubernetes.io/projected/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-kube-api-access-259sk\") pod \"dnsmasq-dns-6584b49599-88dtq\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.657474 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:07:38 crc kubenswrapper[4906]: I1123 07:07:38.719002 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:07:39 crc kubenswrapper[4906]: I1123 07:07:39.175531 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-zhtcl"] Nov 23 07:07:39 crc kubenswrapper[4906]: I1123 07:07:39.184047 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:07:39 crc kubenswrapper[4906]: I1123 07:07:39.237342 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" event={"ID":"c5e77d8e-c265-47e5-ac76-985caeb38679","Type":"ContainerStarted","Data":"abcb659d1602cd71ef96e6416384075e68057590e66b27301822360c78847e0e"} Nov 23 07:07:39 crc kubenswrapper[4906]: I1123 07:07:39.249797 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-88dtq"] Nov 23 07:07:39 crc kubenswrapper[4906]: W1123 07:07:39.250700 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d094e2e_7efa_46df_a4d3_1cd6b3a0bcfa.slice/crio-1af5d66191d103842ab7048dbcb92a27c60d015df1ad02adcf223c9d64d10d55 WatchSource:0}: Error finding container 1af5d66191d103842ab7048dbcb92a27c60d015df1ad02adcf223c9d64d10d55: Status 404 returned error can't find the container with id 1af5d66191d103842ab7048dbcb92a27c60d015df1ad02adcf223c9d64d10d55 Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.247318 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-88dtq" event={"ID":"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa","Type":"ContainerStarted","Data":"1af5d66191d103842ab7048dbcb92a27c60d015df1ad02adcf223c9d64d10d55"} Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.543551 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-88dtq"] Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.573474 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-pf9gv"] Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.574801 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.598871 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-pf9gv"] Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.753542 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-config\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.753872 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.754066 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8lld\" (UniqueName: \"kubernetes.io/projected/8b958001-b533-4f6c-a5d9-e07fd612f484-kube-api-access-w8lld\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.855422 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8lld\" (UniqueName: \"kubernetes.io/projected/8b958001-b533-4f6c-a5d9-e07fd612f484-kube-api-access-w8lld\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.855536 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-config\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.855589 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.856586 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.856867 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-config\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.883307 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8lld\" (UniqueName: \"kubernetes.io/projected/8b958001-b533-4f6c-a5d9-e07fd612f484-kube-api-access-w8lld\") pod \"dnsmasq-dns-7c6d9948dc-pf9gv\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:40 crc kubenswrapper[4906]: I1123 07:07:40.894463 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.209786 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-zhtcl"] Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.248280 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-lqg2k"] Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.249881 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.254088 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-lqg2k"] Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.365366 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-config\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.365440 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-dns-svc\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.365482 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h759h\" (UniqueName: \"kubernetes.io/projected/fcf4d17a-11e9-4282-8883-69d6515799e3-kube-api-access-h759h\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.379033 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-pf9gv"] Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.466509 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-dns-svc\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.467120 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h759h\" (UniqueName: \"kubernetes.io/projected/fcf4d17a-11e9-4282-8883-69d6515799e3-kube-api-access-h759h\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.467211 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-config\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.467716 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-dns-svc\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.467898 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-config\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.499003 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h759h\" (UniqueName: \"kubernetes.io/projected/fcf4d17a-11e9-4282-8883-69d6515799e3-kube-api-access-h759h\") pod \"dnsmasq-dns-6486446b9f-lqg2k\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.573442 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.731313 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.732904 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.740051 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.740370 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.740542 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.740785 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.740950 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.741164 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.741450 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-k7b94" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.763884 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.873793 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a0222ea-7767-4d08-a6ec-6659a33f9df2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874123 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a0222ea-7767-4d08-a6ec-6659a33f9df2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874156 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874178 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874198 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874221 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vpbf\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-kube-api-access-7vpbf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874292 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874327 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874347 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874372 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.874399 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.975759 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.975817 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.975848 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.975899 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.975930 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.976020 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a0222ea-7767-4d08-a6ec-6659a33f9df2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.976046 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a0222ea-7767-4d08-a6ec-6659a33f9df2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.976080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.976108 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.976130 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.976160 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vpbf\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-kube-api-access-7vpbf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.976961 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.978166 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.979375 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.981179 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.981437 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.981735 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.982904 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a0222ea-7767-4d08-a6ec-6659a33f9df2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.983060 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.983332 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a0222ea-7767-4d08-a6ec-6659a33f9df2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:41 crc kubenswrapper[4906]: I1123 07:07:41.987071 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.005125 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vpbf\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-kube-api-access-7vpbf\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.028738 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " pod="openstack/rabbitmq-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.033759 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-lqg2k"] Nov 23 07:07:42 crc kubenswrapper[4906]: W1123 07:07:42.039570 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcf4d17a_11e9_4282_8883_69d6515799e3.slice/crio-2ea3b0ac890f0de980b89cc2345b9317c01acbb3ca6ad073e224953cf720d3a9 WatchSource:0}: Error finding container 2ea3b0ac890f0de980b89cc2345b9317c01acbb3ca6ad073e224953cf720d3a9: Status 404 returned error can't find the container with id 2ea3b0ac890f0de980b89cc2345b9317c01acbb3ca6ad073e224953cf720d3a9 Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.069886 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.265188 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" event={"ID":"fcf4d17a-11e9-4282-8883-69d6515799e3","Type":"ContainerStarted","Data":"2ea3b0ac890f0de980b89cc2345b9317c01acbb3ca6ad073e224953cf720d3a9"} Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.267590 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" event={"ID":"8b958001-b533-4f6c-a5d9-e07fd612f484","Type":"ContainerStarted","Data":"bfec5f53e8791f857245349319e734d9ca3aba74d8519b01a1823271c8d9f116"} Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.307123 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:07:42 crc kubenswrapper[4906]: W1123 07:07:42.318193 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a0222ea_7767_4d08_a6ec_6659a33f9df2.slice/crio-9b2d7b84608f1195fd51adfddfacf4d33692276329ff385853e0c3c2a4c4c088 WatchSource:0}: Error finding container 9b2d7b84608f1195fd51adfddfacf4d33692276329ff385853e0c3c2a4c4c088: Status 404 returned error can't find the container with id 9b2d7b84608f1195fd51adfddfacf4d33692276329ff385853e0c3c2a4c4c088 Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.365504 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.366847 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.369601 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.369792 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.369930 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.370855 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.371097 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.371306 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.371594 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-sh62z" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.383983 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.489541 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.489591 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.489620 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d677a57-1acf-4627-aa66-f4bee96a2b51-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.489672 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.489951 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d677a57-1acf-4627-aa66-f4bee96a2b51-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.490015 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.490056 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.490177 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.490223 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.490270 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.490321 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrvvj\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-kube-api-access-xrvvj\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592063 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592139 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592182 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592213 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrvvj\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-kube-api-access-xrvvj\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592276 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592301 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592332 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d677a57-1acf-4627-aa66-f4bee96a2b51-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592362 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592420 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d677a57-1acf-4627-aa66-f4bee96a2b51-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592459 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592494 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.592912 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.593354 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.593958 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.593990 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.594021 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.594482 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.598250 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.598363 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d677a57-1acf-4627-aa66-f4bee96a2b51-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.600143 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.602226 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d677a57-1acf-4627-aa66-f4bee96a2b51-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.612375 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrvvj\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-kube-api-access-xrvvj\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.619862 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:42 crc kubenswrapper[4906]: I1123 07:07:42.687759 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.160285 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:07:43 crc kubenswrapper[4906]: W1123 07:07:43.171184 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d677a57_1acf_4627_aa66_f4bee96a2b51.slice/crio-3da95091300888aad706cfc8c6e2300a4e5243f39427c92b2cc8b1420752e895 WatchSource:0}: Error finding container 3da95091300888aad706cfc8c6e2300a4e5243f39427c92b2cc8b1420752e895: Status 404 returned error can't find the container with id 3da95091300888aad706cfc8c6e2300a4e5243f39427c92b2cc8b1420752e895 Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.277529 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a0222ea-7767-4d08-a6ec-6659a33f9df2","Type":"ContainerStarted","Data":"9b2d7b84608f1195fd51adfddfacf4d33692276329ff385853e0c3c2a4c4c088"} Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.281093 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d677a57-1acf-4627-aa66-f4bee96a2b51","Type":"ContainerStarted","Data":"3da95091300888aad706cfc8c6e2300a4e5243f39427c92b2cc8b1420752e895"} Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.812241 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.816392 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.817830 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.849212 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.849326 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.849804 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.852537 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gqq6f" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.858532 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917447 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917532 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917574 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-kolla-config\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917606 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-generated\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917640 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-operator-scripts\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917747 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917791 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prddx\" (UniqueName: \"kubernetes.io/projected/adbcc53a-3e0f-47b0-a028-01c686a18205-kube-api-access-prddx\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:43 crc kubenswrapper[4906]: I1123 07:07:43.917851 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-default\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020101 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020159 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prddx\" (UniqueName: \"kubernetes.io/projected/adbcc53a-3e0f-47b0-a028-01c686a18205-kube-api-access-prddx\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020197 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-default\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020237 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020257 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020280 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-kolla-config\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020300 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-generated\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.020323 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-operator-scripts\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.021410 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.021873 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-generated\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.023621 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-default\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.024325 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-operator-scripts\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.025233 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-kolla-config\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.041721 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.064607 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.072370 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prddx\" (UniqueName: \"kubernetes.io/projected/adbcc53a-3e0f-47b0-a028-01c686a18205-kube-api-access-prddx\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.127974 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.174535 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 07:07:44 crc kubenswrapper[4906]: I1123 07:07:44.804969 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 07:07:44 crc kubenswrapper[4906]: W1123 07:07:44.866835 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadbcc53a_3e0f_47b0_a028_01c686a18205.slice/crio-858e0cae8bf896410ba6cada859b5399fbbefca46509879cb8159f22ad27e7f5 WatchSource:0}: Error finding container 858e0cae8bf896410ba6cada859b5399fbbefca46509879cb8159f22ad27e7f5: Status 404 returned error can't find the container with id 858e0cae8bf896410ba6cada859b5399fbbefca46509879cb8159f22ad27e7f5 Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.182190 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.183861 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.192462 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.221343 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.221633 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-7c7lb" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.221816 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.221649 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262385 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262485 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgbzq\" (UniqueName: \"kubernetes.io/projected/0a95e1df-3578-4430-b0ac-befb54723517-kube-api-access-zgbzq\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262511 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0a95e1df-3578-4430-b0ac-befb54723517-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262536 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262564 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262595 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262612 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.262629 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.349213 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adbcc53a-3e0f-47b0-a028-01c686a18205","Type":"ContainerStarted","Data":"858e0cae8bf896410ba6cada859b5399fbbefca46509879cb8159f22ad27e7f5"} Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368555 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgbzq\" (UniqueName: \"kubernetes.io/projected/0a95e1df-3578-4430-b0ac-befb54723517-kube-api-access-zgbzq\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368603 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0a95e1df-3578-4430-b0ac-befb54723517-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368630 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368657 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368704 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368725 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368743 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.368765 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.369489 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0a95e1df-3578-4430-b0ac-befb54723517-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.373761 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.381886 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.386535 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.386778 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.387033 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.391241 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.393315 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.395950 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgbzq\" (UniqueName: \"kubernetes.io/projected/0a95e1df-3578-4430-b0ac-befb54723517-kube-api-access-zgbzq\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.400356 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.407310 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.412809 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.413892 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.424963 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.425363 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-sq6pz" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.425533 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.473341 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.479201 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kolla-config\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.479306 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-config-data\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.479467 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.479555 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.479839 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2928c\" (UniqueName: \"kubernetes.io/projected/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kube-api-access-2928c\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.503388 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.567093 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-7c7lb" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.583947 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.605273 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.605372 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.605392 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2928c\" (UniqueName: \"kubernetes.io/projected/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kube-api-access-2928c\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.605476 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kolla-config\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.605525 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-config-data\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.606376 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-config-data\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.607329 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kolla-config\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.618025 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.621222 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.635216 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2928c\" (UniqueName: \"kubernetes.io/projected/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kube-api-access-2928c\") pod \"memcached-0\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " pod="openstack/memcached-0" Nov 23 07:07:45 crc kubenswrapper[4906]: I1123 07:07:45.826072 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 07:07:46 crc kubenswrapper[4906]: I1123 07:07:46.164241 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 07:07:46 crc kubenswrapper[4906]: I1123 07:07:46.169817 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 23 07:07:46 crc kubenswrapper[4906]: W1123 07:07:46.194428 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27d994c3_4d19_4084_8c09_d0f2a8d7ede7.slice/crio-8527db811030ccea75981691598f1f7afdfe79a7d8eb1407e57d34e3f5276308 WatchSource:0}: Error finding container 8527db811030ccea75981691598f1f7afdfe79a7d8eb1407e57d34e3f5276308: Status 404 returned error can't find the container with id 8527db811030ccea75981691598f1f7afdfe79a7d8eb1407e57d34e3f5276308 Nov 23 07:07:46 crc kubenswrapper[4906]: W1123 07:07:46.195494 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a95e1df_3578_4430_b0ac_befb54723517.slice/crio-b27a842c89676241116fe33499f637397fd007b238dd4c2a6f1c1470abbc6662 WatchSource:0}: Error finding container b27a842c89676241116fe33499f637397fd007b238dd4c2a6f1c1470abbc6662: Status 404 returned error can't find the container with id b27a842c89676241116fe33499f637397fd007b238dd4c2a6f1c1470abbc6662 Nov 23 07:07:46 crc kubenswrapper[4906]: I1123 07:07:46.384588 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0a95e1df-3578-4430-b0ac-befb54723517","Type":"ContainerStarted","Data":"b27a842c89676241116fe33499f637397fd007b238dd4c2a6f1c1470abbc6662"} Nov 23 07:07:46 crc kubenswrapper[4906]: I1123 07:07:46.391294 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"27d994c3-4d19-4084-8c09-d0f2a8d7ede7","Type":"ContainerStarted","Data":"8527db811030ccea75981691598f1f7afdfe79a7d8eb1407e57d34e3f5276308"} Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.072863 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.075965 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.085791 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-lvw9v" Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.108817 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.146728 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f25fh\" (UniqueName: \"kubernetes.io/projected/e368b3d0-8ac9-49be-b2c4-8e88d8327811-kube-api-access-f25fh\") pod \"kube-state-metrics-0\" (UID: \"e368b3d0-8ac9-49be-b2c4-8e88d8327811\") " pod="openstack/kube-state-metrics-0" Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.254629 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f25fh\" (UniqueName: \"kubernetes.io/projected/e368b3d0-8ac9-49be-b2c4-8e88d8327811-kube-api-access-f25fh\") pod \"kube-state-metrics-0\" (UID: \"e368b3d0-8ac9-49be-b2c4-8e88d8327811\") " pod="openstack/kube-state-metrics-0" Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.278537 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f25fh\" (UniqueName: \"kubernetes.io/projected/e368b3d0-8ac9-49be-b2c4-8e88d8327811-kube-api-access-f25fh\") pod \"kube-state-metrics-0\" (UID: \"e368b3d0-8ac9-49be-b2c4-8e88d8327811\") " pod="openstack/kube-state-metrics-0" Nov 23 07:07:47 crc kubenswrapper[4906]: I1123 07:07:47.407040 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:07:48 crc kubenswrapper[4906]: I1123 07:07:48.091455 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:07:48 crc kubenswrapper[4906]: I1123 07:07:48.549496 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e368b3d0-8ac9-49be-b2c4-8e88d8327811","Type":"ContainerStarted","Data":"aa2feed880b147f6869bba98d26de9ee9ad67b63a10e56e9017b6a965cd060de"} Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.400312 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l9d4x"] Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.405107 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.409117 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-g8pp2" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.410864 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l9d4x"] Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.411060 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.439492 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-ftw6h"] Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.441214 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.444792 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.454117 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67qwj\" (UniqueName: \"kubernetes.io/projected/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-kube-api-access-67qwj\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.454367 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-log-ovn\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.454443 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run-ovn\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.454557 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.454733 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-scripts\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.454801 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-ovn-controller-tls-certs\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.454920 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-combined-ca-bundle\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.460174 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ftw6h"] Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556583 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run-ovn\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556647 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-lib\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556692 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556734 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-scripts\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556784 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-ovn-controller-tls-certs\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556800 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-log\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556825 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-run\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556855 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-combined-ca-bundle\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556874 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-etc-ovs\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556892 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r4td\" (UniqueName: \"kubernetes.io/projected/e34ddce9-d32e-4c31-9716-86ad39a590e9-kube-api-access-6r4td\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556921 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67qwj\" (UniqueName: \"kubernetes.io/projected/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-kube-api-access-67qwj\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556941 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e34ddce9-d32e-4c31-9716-86ad39a590e9-scripts\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.556982 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-log-ovn\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.559556 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-scripts\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.561118 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run-ovn\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.561381 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-log-ovn\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.562485 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.565105 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-ovn-controller-tls-certs\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.578629 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67qwj\" (UniqueName: \"kubernetes.io/projected/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-kube-api-access-67qwj\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.585338 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-combined-ca-bundle\") pod \"ovn-controller-l9d4x\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658128 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-lib\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658221 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-log\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658246 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-run\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658272 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-etc-ovs\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658296 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r4td\" (UniqueName: \"kubernetes.io/projected/e34ddce9-d32e-4c31-9716-86ad39a590e9-kube-api-access-6r4td\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658349 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e34ddce9-d32e-4c31-9716-86ad39a590e9-scripts\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658455 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-log\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658459 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-lib\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658715 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-etc-ovs\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.658852 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-run\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.661586 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e34ddce9-d32e-4c31-9716-86ad39a590e9-scripts\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.679534 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r4td\" (UniqueName: \"kubernetes.io/projected/e34ddce9-d32e-4c31-9716-86ad39a590e9-kube-api-access-6r4td\") pod \"ovn-controller-ovs-ftw6h\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.739813 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x" Nov 23 07:07:51 crc kubenswrapper[4906]: I1123 07:07:51.772936 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.706044 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.710656 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.714640 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.716216 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vj749" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.716376 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.716271 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.716271 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.721987 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.821761 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pk52\" (UniqueName: \"kubernetes.io/projected/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-kube-api-access-5pk52\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.821825 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.821870 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.821953 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-config\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.821981 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.822168 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.822426 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.822451 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.893573 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.902931 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.905478 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.907945 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.909300 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.909430 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.909465 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-cj52c" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.927894 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.927955 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.927993 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pk52\" (UniqueName: \"kubernetes.io/projected/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-kube-api-access-5pk52\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.928027 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.928068 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.928090 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-config\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.928120 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.928151 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.928586 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.929317 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.932990 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-config\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.933018 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.935974 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.936360 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.938289 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.952178 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pk52\" (UniqueName: \"kubernetes.io/projected/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-kube-api-access-5pk52\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:54 crc kubenswrapper[4906]: I1123 07:07:54.958322 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030136 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030195 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030228 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030263 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030279 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-config\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030306 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030341 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5zkc\" (UniqueName: \"kubernetes.io/projected/51dd433b-f1a2-422a-bf5b-643e150c3bca-kube-api-access-r5zkc\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.030403 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.043879 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132226 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132302 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132329 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132368 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132403 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132424 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-config\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132457 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132497 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5zkc\" (UniqueName: \"kubernetes.io/projected/51dd433b-f1a2-422a-bf5b-643e150c3bca-kube-api-access-r5zkc\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.132697 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.133290 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.133982 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.133991 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-config\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.145429 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.147706 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.154372 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.155671 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5zkc\" (UniqueName: \"kubernetes.io/projected/51dd433b-f1a2-422a-bf5b-643e150c3bca-kube-api-access-r5zkc\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.164172 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " pod="openstack/ovsdbserver-nb-0" Nov 23 07:07:55 crc kubenswrapper[4906]: I1123 07:07:55.221812 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.504122 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.505049 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xrvvj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(4d677a57-1acf-4627-aa66-f4bee96a2b51): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.506236 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.525932 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.526143 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7vpbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(9a0222ea-7767-4d08-a6ec-6659a33f9df2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.527306 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.706267 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b\\\"\"" pod="openstack/rabbitmq-server-0" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" Nov 23 07:08:03 crc kubenswrapper[4906]: E1123 07:08:03.706526 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.098731 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.099756 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ppccf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bdd77c89-zhtcl_openstack(c5e77d8e-c265-47e5-ac76-985caeb38679): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.101430 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" podUID="c5e77d8e-c265-47e5-ac76-985caeb38679" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.124256 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.124496 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h759h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6486446b9f-lqg2k_openstack(fcf4d17a-11e9-4282-8883-69d6515799e3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.124867 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.124965 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-259sk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6584b49599-88dtq_openstack(0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.126308 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6584b49599-88dtq" podUID="0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.126413 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" podUID="fcf4d17a-11e9-4282-8883-69d6515799e3" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.140769 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.140970 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w8lld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7c6d9948dc-pf9gv_openstack(8b958001-b533-4f6c-a5d9-e07fd612f484): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.142162 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" podUID="8b958001-b533-4f6c-a5d9-e07fd612f484" Nov 23 07:08:12 crc kubenswrapper[4906]: I1123 07:08:12.571859 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l9d4x"] Nov 23 07:08:12 crc kubenswrapper[4906]: I1123 07:08:12.649114 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 07:08:12 crc kubenswrapper[4906]: I1123 07:08:12.747072 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.778987 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" podUID="fcf4d17a-11e9-4282-8883-69d6515799e3" Nov 23 07:08:12 crc kubenswrapper[4906]: E1123 07:08:12.779440 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" podUID="8b958001-b533-4f6c-a5d9-e07fd612f484" Nov 23 07:08:12 crc kubenswrapper[4906]: I1123 07:08:12.896874 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ftw6h"] Nov 23 07:08:13 crc kubenswrapper[4906]: W1123 07:08:13.173164 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51dd433b_f1a2_422a_bf5b_643e150c3bca.slice/crio-56e048bf7af7ca4d54382a670d2c0118e5ad996cb59a2692751315f09133d3a8 WatchSource:0}: Error finding container 56e048bf7af7ca4d54382a670d2c0118e5ad996cb59a2692751315f09133d3a8: Status 404 returned error can't find the container with id 56e048bf7af7ca4d54382a670d2c0118e5ad996cb59a2692751315f09133d3a8 Nov 23 07:08:13 crc kubenswrapper[4906]: W1123 07:08:13.175099 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ed394c7_5cf6_4a7b_8060_6ffb4953fbfd.slice/crio-031a9b7cf5969785a11c06115f5c7975fef7a3f132b1990849b16ed8da9f3345 WatchSource:0}: Error finding container 031a9b7cf5969785a11c06115f5c7975fef7a3f132b1990849b16ed8da9f3345: Status 404 returned error can't find the container with id 031a9b7cf5969785a11c06115f5c7975fef7a3f132b1990849b16ed8da9f3345 Nov 23 07:08:13 crc kubenswrapper[4906]: W1123 07:08:13.182217 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd96cfa1a_48d0_4efd_8594_4386a5d7ef35.slice/crio-4dab160dbcb6dc297117e7a0680bd44a1741500bb00c62e450d050c5c4b6ccb3 WatchSource:0}: Error finding container 4dab160dbcb6dc297117e7a0680bd44a1741500bb00c62e450d050c5c4b6ccb3: Status 404 returned error can't find the container with id 4dab160dbcb6dc297117e7a0680bd44a1741500bb00c62e450d050c5c4b6ccb3 Nov 23 07:08:13 crc kubenswrapper[4906]: E1123 07:08:13.311846 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Nov 23 07:08:13 crc kubenswrapper[4906]: E1123 07:08:13.311923 4906 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Nov 23 07:08:13 crc kubenswrapper[4906]: E1123 07:08:13.314307 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f25fh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(e368b3d0-8ac9-49be-b2c4-8e88d8327811): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 07:08:13 crc kubenswrapper[4906]: E1123 07:08:13.315532 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.465624 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.503153 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.617604 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-dns-svc\") pod \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.617699 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5e77d8e-c265-47e5-ac76-985caeb38679-config\") pod \"c5e77d8e-c265-47e5-ac76-985caeb38679\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.617909 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppccf\" (UniqueName: \"kubernetes.io/projected/c5e77d8e-c265-47e5-ac76-985caeb38679-kube-api-access-ppccf\") pod \"c5e77d8e-c265-47e5-ac76-985caeb38679\" (UID: \"c5e77d8e-c265-47e5-ac76-985caeb38679\") " Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.618034 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-259sk\" (UniqueName: \"kubernetes.io/projected/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-kube-api-access-259sk\") pod \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.618095 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-config\") pod \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\" (UID: \"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa\") " Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.619586 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5e77d8e-c265-47e5-ac76-985caeb38679-config" (OuterVolumeSpecName: "config") pod "c5e77d8e-c265-47e5-ac76-985caeb38679" (UID: "c5e77d8e-c265-47e5-ac76-985caeb38679"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.619599 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa" (UID: "0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.619924 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-config" (OuterVolumeSpecName: "config") pod "0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa" (UID: "0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.626116 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-kube-api-access-259sk" (OuterVolumeSpecName: "kube-api-access-259sk") pod "0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa" (UID: "0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa"). InnerVolumeSpecName "kube-api-access-259sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.627868 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5e77d8e-c265-47e5-ac76-985caeb38679-kube-api-access-ppccf" (OuterVolumeSpecName: "kube-api-access-ppccf") pod "c5e77d8e-c265-47e5-ac76-985caeb38679" (UID: "c5e77d8e-c265-47e5-ac76-985caeb38679"). InnerVolumeSpecName "kube-api-access-ppccf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.719808 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppccf\" (UniqueName: \"kubernetes.io/projected/c5e77d8e-c265-47e5-ac76-985caeb38679-kube-api-access-ppccf\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.719842 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-259sk\" (UniqueName: \"kubernetes.io/projected/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-kube-api-access-259sk\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.719853 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.719863 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.719872 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5e77d8e-c265-47e5-ac76-985caeb38679-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.790600 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x" event={"ID":"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd","Type":"ContainerStarted","Data":"031a9b7cf5969785a11c06115f5c7975fef7a3f132b1990849b16ed8da9f3345"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.792740 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" event={"ID":"c5e77d8e-c265-47e5-ac76-985caeb38679","Type":"ContainerDied","Data":"abcb659d1602cd71ef96e6416384075e68057590e66b27301822360c78847e0e"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.792772 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-zhtcl" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.795110 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"51dd433b-f1a2-422a-bf5b-643e150c3bca","Type":"ContainerStarted","Data":"56e048bf7af7ca4d54382a670d2c0118e5ad996cb59a2692751315f09133d3a8"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.798319 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-88dtq" event={"ID":"0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa","Type":"ContainerDied","Data":"1af5d66191d103842ab7048dbcb92a27c60d015df1ad02adcf223c9d64d10d55"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.798407 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-88dtq" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.801779 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d96cfa1a-48d0-4efd-8594-4386a5d7ef35","Type":"ContainerStarted","Data":"4dab160dbcb6dc297117e7a0680bd44a1741500bb00c62e450d050c5c4b6ccb3"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.805102 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ftw6h" event={"ID":"e34ddce9-d32e-4c31-9716-86ad39a590e9","Type":"ContainerStarted","Data":"905ba1126072f9b90864ddc6b6599d63d7f49861d6bb3ab0ed2379c179665006"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.806789 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adbcc53a-3e0f-47b0-a028-01c686a18205","Type":"ContainerStarted","Data":"4e5900f7a006c98c9c587b8cb875127576a79a2f875e56f9d5bcbf31a06d3004"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.810254 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0a95e1df-3578-4430-b0ac-befb54723517","Type":"ContainerStarted","Data":"ed942ae5cec5564c89129803c0336b09e9319da12fa2443229288d6c827463c8"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.815268 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"27d994c3-4d19-4084-8c09-d0f2a8d7ede7","Type":"ContainerStarted","Data":"c62d53e7d032049f41cdae0eac7ed993f6453996187bc0ab5a5c3458db844d22"} Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.815327 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 23 07:08:13 crc kubenswrapper[4906]: E1123 07:08:13.816827 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.868990 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.9795047869999998 podStartE2EDuration="28.868971097s" podCreationTimestamp="2025-11-23 07:07:45 +0000 UTC" firstStartedPulling="2025-11-23 07:07:46.203880496 +0000 UTC m=+1081.717271789" lastFinishedPulling="2025-11-23 07:08:11.093346786 +0000 UTC m=+1106.606738099" observedRunningTime="2025-11-23 07:08:13.862946559 +0000 UTC m=+1109.376337862" watchObservedRunningTime="2025-11-23 07:08:13.868971097 +0000 UTC m=+1109.382362390" Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.950073 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-zhtcl"] Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.956942 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-zhtcl"] Nov 23 07:08:13 crc kubenswrapper[4906]: I1123 07:08:13.998184 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-88dtq"] Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.003907 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-88dtq"] Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.549944 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-njmjc"] Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.554205 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.556897 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.564065 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-njmjc"] Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.645892 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.645970 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovn-rundir\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.645997 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88cqt\" (UniqueName: \"kubernetes.io/projected/3acf9ff7-0012-4e1a-b940-037bebcc4c88-kube-api-access-88cqt\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.646023 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3acf9ff7-0012-4e1a-b940-037bebcc4c88-config\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.646048 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-combined-ca-bundle\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.646089 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovs-rundir\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748169 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748283 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovn-rundir\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748308 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88cqt\" (UniqueName: \"kubernetes.io/projected/3acf9ff7-0012-4e1a-b940-037bebcc4c88-kube-api-access-88cqt\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748343 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3acf9ff7-0012-4e1a-b940-037bebcc4c88-config\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748370 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-combined-ca-bundle\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748410 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovs-rundir\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748795 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovs-rundir\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.748865 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovn-rundir\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.750105 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3acf9ff7-0012-4e1a-b940-037bebcc4c88-config\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.759383 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-combined-ca-bundle\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.765656 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88cqt\" (UniqueName: \"kubernetes.io/projected/3acf9ff7-0012-4e1a-b940-037bebcc4c88-kube-api-access-88cqt\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.766484 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-njmjc\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.901425 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-pf9gv"] Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.919649 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-xgt22"] Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.924607 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.929020 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.943371 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 23 07:08:14 crc kubenswrapper[4906]: I1123 07:08:14.972410 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-xgt22"] Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.054569 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlgsg\" (UniqueName: \"kubernetes.io/projected/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-kube-api-access-zlgsg\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.054666 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-config\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.055053 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.055118 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.130934 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-lqg2k"] Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.146613 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-hdlmk"] Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.148373 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.151794 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.152543 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-hdlmk"] Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.156613 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.156660 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.156822 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlgsg\" (UniqueName: \"kubernetes.io/projected/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-kube-api-access-zlgsg\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.156893 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-config\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.159399 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-config\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.159723 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.160420 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.176311 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlgsg\" (UniqueName: \"kubernetes.io/projected/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-kube-api-access-zlgsg\") pod \"dnsmasq-dns-65c9b8d4f7-xgt22\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.261075 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.262749 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.262779 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.262795 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.262844 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpk5b\" (UniqueName: \"kubernetes.io/projected/38c90719-7ad4-4864-ab04-fe94caecd409-kube-api-access-cpk5b\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.262903 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-config\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.364273 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.364327 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.364367 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.364433 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpk5b\" (UniqueName: \"kubernetes.io/projected/38c90719-7ad4-4864-ab04-fe94caecd409-kube-api-access-cpk5b\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.364518 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-config\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.366226 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-config\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.367315 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.369015 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.371839 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.375168 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa" path="/var/lib/kubelet/pods/0d094e2e-7efa-46df-a4d3-1cd6b3a0bcfa/volumes" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.375552 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5e77d8e-c265-47e5-ac76-985caeb38679" path="/var/lib/kubelet/pods/c5e77d8e-c265-47e5-ac76-985caeb38679/volumes" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.433431 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpk5b\" (UniqueName: \"kubernetes.io/projected/38c90719-7ad4-4864-ab04-fe94caecd409-kube-api-access-cpk5b\") pod \"dnsmasq-dns-5c476d78c5-hdlmk\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.440479 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.571812 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-dns-svc\") pod \"8b958001-b533-4f6c-a5d9-e07fd612f484\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.572418 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8lld\" (UniqueName: \"kubernetes.io/projected/8b958001-b533-4f6c-a5d9-e07fd612f484-kube-api-access-w8lld\") pod \"8b958001-b533-4f6c-a5d9-e07fd612f484\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.572461 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-config\") pod \"8b958001-b533-4f6c-a5d9-e07fd612f484\" (UID: \"8b958001-b533-4f6c-a5d9-e07fd612f484\") " Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.573932 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-config" (OuterVolumeSpecName: "config") pod "8b958001-b533-4f6c-a5d9-e07fd612f484" (UID: "8b958001-b533-4f6c-a5d9-e07fd612f484"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.574127 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8b958001-b533-4f6c-a5d9-e07fd612f484" (UID: "8b958001-b533-4f6c-a5d9-e07fd612f484"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.579612 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b958001-b533-4f6c-a5d9-e07fd612f484-kube-api-access-w8lld" (OuterVolumeSpecName: "kube-api-access-w8lld") pod "8b958001-b533-4f6c-a5d9-e07fd612f484" (UID: "8b958001-b533-4f6c-a5d9-e07fd612f484"). InnerVolumeSpecName "kube-api-access-w8lld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.581905 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.626409 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.674497 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h759h\" (UniqueName: \"kubernetes.io/projected/fcf4d17a-11e9-4282-8883-69d6515799e3-kube-api-access-h759h\") pod \"fcf4d17a-11e9-4282-8883-69d6515799e3\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.674556 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-dns-svc\") pod \"fcf4d17a-11e9-4282-8883-69d6515799e3\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.674784 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-config\") pod \"fcf4d17a-11e9-4282-8883-69d6515799e3\" (UID: \"fcf4d17a-11e9-4282-8883-69d6515799e3\") " Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.675364 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8lld\" (UniqueName: \"kubernetes.io/projected/8b958001-b533-4f6c-a5d9-e07fd612f484-kube-api-access-w8lld\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.675352 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fcf4d17a-11e9-4282-8883-69d6515799e3" (UID: "fcf4d17a-11e9-4282-8883-69d6515799e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.675383 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.675445 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b958001-b533-4f6c-a5d9-e07fd612f484-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.675694 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-config" (OuterVolumeSpecName: "config") pod "fcf4d17a-11e9-4282-8883-69d6515799e3" (UID: "fcf4d17a-11e9-4282-8883-69d6515799e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.693515 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcf4d17a-11e9-4282-8883-69d6515799e3-kube-api-access-h759h" (OuterVolumeSpecName: "kube-api-access-h759h") pod "fcf4d17a-11e9-4282-8883-69d6515799e3" (UID: "fcf4d17a-11e9-4282-8883-69d6515799e3"). InnerVolumeSpecName "kube-api-access-h759h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.777326 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h759h\" (UniqueName: \"kubernetes.io/projected/fcf4d17a-11e9-4282-8883-69d6515799e3-kube-api-access-h759h\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.777355 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.777365 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf4d17a-11e9-4282-8883-69d6515799e3-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.792252 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-njmjc"] Nov 23 07:08:15 crc kubenswrapper[4906]: W1123 07:08:15.814140 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3acf9ff7_0012_4e1a_b940_037bebcc4c88.slice/crio-b1f9d156bc38ee618a38c1d67bd224d7161e43b996c174f5ccf1865ea86ef638 WatchSource:0}: Error finding container b1f9d156bc38ee618a38c1d67bd224d7161e43b996c174f5ccf1865ea86ef638: Status 404 returned error can't find the container with id b1f9d156bc38ee618a38c1d67bd224d7161e43b996c174f5ccf1865ea86ef638 Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.851606 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" event={"ID":"fcf4d17a-11e9-4282-8883-69d6515799e3","Type":"ContainerDied","Data":"2ea3b0ac890f0de980b89cc2345b9317c01acbb3ca6ad073e224953cf720d3a9"} Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.851642 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-lqg2k" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.852832 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-njmjc" event={"ID":"3acf9ff7-0012-4e1a-b940-037bebcc4c88","Type":"ContainerStarted","Data":"b1f9d156bc38ee618a38c1d67bd224d7161e43b996c174f5ccf1865ea86ef638"} Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.857917 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" event={"ID":"8b958001-b533-4f6c-a5d9-e07fd612f484","Type":"ContainerDied","Data":"bfec5f53e8791f857245349319e734d9ca3aba74d8519b01a1823271c8d9f116"} Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.858073 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-pf9gv" Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.873333 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-xgt22"] Nov 23 07:08:15 crc kubenswrapper[4906]: W1123 07:08:15.889959 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod607acb77_3bfa_4e7c_8aef_07597f2f2e1a.slice/crio-1bca03efe4961b4a867c44cefc65931ca7576266356308eb273b4411ba7ef572 WatchSource:0}: Error finding container 1bca03efe4961b4a867c44cefc65931ca7576266356308eb273b4411ba7ef572: Status 404 returned error can't find the container with id 1bca03efe4961b4a867c44cefc65931ca7576266356308eb273b4411ba7ef572 Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.936767 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-pf9gv"] Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.953286 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-pf9gv"] Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.969704 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-lqg2k"] Nov 23 07:08:15 crc kubenswrapper[4906]: I1123 07:08:15.980219 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-lqg2k"] Nov 23 07:08:16 crc kubenswrapper[4906]: I1123 07:08:16.123262 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-hdlmk"] Nov 23 07:08:16 crc kubenswrapper[4906]: I1123 07:08:16.868070 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" event={"ID":"607acb77-3bfa-4e7c-8aef-07597f2f2e1a","Type":"ContainerStarted","Data":"1bca03efe4961b4a867c44cefc65931ca7576266356308eb273b4411ba7ef572"} Nov 23 07:08:16 crc kubenswrapper[4906]: I1123 07:08:16.869862 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a0222ea-7767-4d08-a6ec-6659a33f9df2","Type":"ContainerStarted","Data":"e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44"} Nov 23 07:08:17 crc kubenswrapper[4906]: I1123 07:08:17.377434 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b958001-b533-4f6c-a5d9-e07fd612f484" path="/var/lib/kubelet/pods/8b958001-b533-4f6c-a5d9-e07fd612f484/volumes" Nov 23 07:08:17 crc kubenswrapper[4906]: I1123 07:08:17.377830 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcf4d17a-11e9-4282-8883-69d6515799e3" path="/var/lib/kubelet/pods/fcf4d17a-11e9-4282-8883-69d6515799e3/volumes" Nov 23 07:08:17 crc kubenswrapper[4906]: I1123 07:08:17.881376 4906 generic.go:334] "Generic (PLEG): container finished" podID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerID="4e5900f7a006c98c9c587b8cb875127576a79a2f875e56f9d5bcbf31a06d3004" exitCode=0 Nov 23 07:08:17 crc kubenswrapper[4906]: I1123 07:08:17.881509 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adbcc53a-3e0f-47b0-a028-01c686a18205","Type":"ContainerDied","Data":"4e5900f7a006c98c9c587b8cb875127576a79a2f875e56f9d5bcbf31a06d3004"} Nov 23 07:08:17 crc kubenswrapper[4906]: I1123 07:08:17.885694 4906 generic.go:334] "Generic (PLEG): container finished" podID="0a95e1df-3578-4430-b0ac-befb54723517" containerID="ed942ae5cec5564c89129803c0336b09e9319da12fa2443229288d6c827463c8" exitCode=0 Nov 23 07:08:17 crc kubenswrapper[4906]: I1123 07:08:17.885731 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0a95e1df-3578-4430-b0ac-befb54723517","Type":"ContainerDied","Data":"ed942ae5cec5564c89129803c0336b09e9319da12fa2443229288d6c827463c8"} Nov 23 07:08:18 crc kubenswrapper[4906]: W1123 07:08:18.070249 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38c90719_7ad4_4864_ab04_fe94caecd409.slice/crio-472ba886c332b94bf313b14c820fadb76102ef49ba60fa0d3f58862847fb3f3d WatchSource:0}: Error finding container 472ba886c332b94bf313b14c820fadb76102ef49ba60fa0d3f58862847fb3f3d: Status 404 returned error can't find the container with id 472ba886c332b94bf313b14c820fadb76102ef49ba60fa0d3f58862847fb3f3d Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.907330 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0a95e1df-3578-4430-b0ac-befb54723517","Type":"ContainerStarted","Data":"8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85"} Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.914040 4906 generic.go:334] "Generic (PLEG): container finished" podID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerID="b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668" exitCode=0 Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.914145 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" event={"ID":"607acb77-3bfa-4e7c-8aef-07597f2f2e1a","Type":"ContainerDied","Data":"b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668"} Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.919862 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d96cfa1a-48d0-4efd-8594-4386a5d7ef35","Type":"ContainerStarted","Data":"c10b96182b460a717b72550d27ff7aae433f2b3f9e4bf71a704dc8b41e9653fb"} Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.921662 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"51dd433b-f1a2-422a-bf5b-643e150c3bca","Type":"ContainerStarted","Data":"2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0"} Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.930971 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" event={"ID":"38c90719-7ad4-4864-ab04-fe94caecd409","Type":"ContainerStarted","Data":"472ba886c332b94bf313b14c820fadb76102ef49ba60fa0d3f58862847fb3f3d"} Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.945344 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adbcc53a-3e0f-47b0-a028-01c686a18205","Type":"ContainerStarted","Data":"27497e40b508a21677f03a6794d604b1ad821f2140f70be90b3a8b007d2d719d"} Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.956480 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=9.15951306 podStartE2EDuration="34.95645143s" podCreationTimestamp="2025-11-23 07:07:44 +0000 UTC" firstStartedPulling="2025-11-23 07:07:46.219968399 +0000 UTC m=+1081.733359702" lastFinishedPulling="2025-11-23 07:08:12.016906729 +0000 UTC m=+1107.530298072" observedRunningTime="2025-11-23 07:08:18.929081505 +0000 UTC m=+1114.442472808" watchObservedRunningTime="2025-11-23 07:08:18.95645143 +0000 UTC m=+1114.469842743" Nov 23 07:08:18 crc kubenswrapper[4906]: I1123 07:08:18.981319 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.783326596 podStartE2EDuration="36.981294248s" podCreationTimestamp="2025-11-23 07:07:42 +0000 UTC" firstStartedPulling="2025-11-23 07:07:44.892265312 +0000 UTC m=+1080.405656615" lastFinishedPulling="2025-11-23 07:08:11.090232924 +0000 UTC m=+1106.603624267" observedRunningTime="2025-11-23 07:08:18.976886703 +0000 UTC m=+1114.490278006" watchObservedRunningTime="2025-11-23 07:08:18.981294248 +0000 UTC m=+1114.494685551" Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.960197 4906 generic.go:334] "Generic (PLEG): container finished" podID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerID="85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0" exitCode=0 Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.960316 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ftw6h" event={"ID":"e34ddce9-d32e-4c31-9716-86ad39a590e9","Type":"ContainerDied","Data":"85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0"} Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.966026 4906 generic.go:334] "Generic (PLEG): container finished" podID="38c90719-7ad4-4864-ab04-fe94caecd409" containerID="c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625" exitCode=0 Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.966121 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" event={"ID":"38c90719-7ad4-4864-ab04-fe94caecd409","Type":"ContainerDied","Data":"c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625"} Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.971588 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d677a57-1acf-4627-aa66-f4bee96a2b51","Type":"ContainerStarted","Data":"89f5735b86ee117a6dd64d5fbfd44ffb2a0e87dd340dc5bc635c3b876a04901a"} Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.976505 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" event={"ID":"607acb77-3bfa-4e7c-8aef-07597f2f2e1a","Type":"ContainerStarted","Data":"34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44"} Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.977277 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.983142 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x" event={"ID":"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd","Type":"ContainerStarted","Data":"5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e"} Nov 23 07:08:19 crc kubenswrapper[4906]: I1123 07:08:19.984916 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-l9d4x" Nov 23 07:08:20 crc kubenswrapper[4906]: I1123 07:08:20.026958 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" podStartSLOduration=3.526838433 podStartE2EDuration="6.02693815s" podCreationTimestamp="2025-11-23 07:08:14 +0000 UTC" firstStartedPulling="2025-11-23 07:08:15.894667557 +0000 UTC m=+1111.408058860" lastFinishedPulling="2025-11-23 07:08:18.394767274 +0000 UTC m=+1113.908158577" observedRunningTime="2025-11-23 07:08:20.022670748 +0000 UTC m=+1115.536062051" watchObservedRunningTime="2025-11-23 07:08:20.02693815 +0000 UTC m=+1115.540329443" Nov 23 07:08:20 crc kubenswrapper[4906]: I1123 07:08:20.063402 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-l9d4x" podStartSLOduration=23.790801016 podStartE2EDuration="29.063378351s" podCreationTimestamp="2025-11-23 07:07:51 +0000 UTC" firstStartedPulling="2025-11-23 07:08:13.180756418 +0000 UTC m=+1108.694147721" lastFinishedPulling="2025-11-23 07:08:18.453333753 +0000 UTC m=+1113.966725056" observedRunningTime="2025-11-23 07:08:20.05835881 +0000 UTC m=+1115.571750113" watchObservedRunningTime="2025-11-23 07:08:20.063378351 +0000 UTC m=+1115.576769654" Nov 23 07:08:20 crc kubenswrapper[4906]: I1123 07:08:20.827864 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 23 07:08:20 crc kubenswrapper[4906]: E1123 07:08:20.983966 4906 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.227:37640->38.102.83.227:35181: write tcp 38.102.83.227:37640->38.102.83.227:35181: write: broken pipe Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.003157 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-njmjc" event={"ID":"3acf9ff7-0012-4e1a-b940-037bebcc4c88","Type":"ContainerStarted","Data":"28d9044f3cc3149e597f205efe1d60660461a96061a5b3c9310711ba4008ce47"} Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.016406 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d96cfa1a-48d0-4efd-8594-4386a5d7ef35","Type":"ContainerStarted","Data":"a901c2249d99c2e7c335898965028b8ed5f7a9b34574ccbb6be133b51490d695"} Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.021593 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-njmjc" podStartSLOduration=2.294567889 podStartE2EDuration="8.02158053s" podCreationTimestamp="2025-11-23 07:08:14 +0000 UTC" firstStartedPulling="2025-11-23 07:08:15.817277627 +0000 UTC m=+1111.330668930" lastFinishedPulling="2025-11-23 07:08:21.544290268 +0000 UTC m=+1117.057681571" observedRunningTime="2025-11-23 07:08:22.021157258 +0000 UTC m=+1117.534548571" watchObservedRunningTime="2025-11-23 07:08:22.02158053 +0000 UTC m=+1117.534971833" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.023253 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"51dd433b-f1a2-422a-bf5b-643e150c3bca","Type":"ContainerStarted","Data":"a091f675900348c0b86ab0b4409f0e26e962032b4b02dd301de38636bc1e7e50"} Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.036717 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ftw6h" event={"ID":"e34ddce9-d32e-4c31-9716-86ad39a590e9","Type":"ContainerStarted","Data":"cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4"} Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.041768 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" event={"ID":"38c90719-7ad4-4864-ab04-fe94caecd409","Type":"ContainerStarted","Data":"dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16"} Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.041806 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.047823 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.079868 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.747846384 podStartE2EDuration="29.07984067s" podCreationTimestamp="2025-11-23 07:07:53 +0000 UTC" firstStartedPulling="2025-11-23 07:08:13.195115193 +0000 UTC m=+1108.708506486" lastFinishedPulling="2025-11-23 07:08:21.527109469 +0000 UTC m=+1117.040500772" observedRunningTime="2025-11-23 07:08:22.05988989 +0000 UTC m=+1117.573281193" watchObservedRunningTime="2025-11-23 07:08:22.07984067 +0000 UTC m=+1117.593231963" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.088488 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=20.745138774 podStartE2EDuration="29.088464046s" podCreationTimestamp="2025-11-23 07:07:53 +0000 UTC" firstStartedPulling="2025-11-23 07:08:13.17894332 +0000 UTC m=+1108.692334623" lastFinishedPulling="2025-11-23 07:08:21.522268592 +0000 UTC m=+1117.035659895" observedRunningTime="2025-11-23 07:08:22.076796811 +0000 UTC m=+1117.590188114" watchObservedRunningTime="2025-11-23 07:08:22.088464046 +0000 UTC m=+1117.601855349" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.129606 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.130438 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" podStartSLOduration=6.570150653 podStartE2EDuration="7.130404091s" podCreationTimestamp="2025-11-23 07:08:15 +0000 UTC" firstStartedPulling="2025-11-23 07:08:18.09317686 +0000 UTC m=+1113.606568163" lastFinishedPulling="2025-11-23 07:08:18.653430298 +0000 UTC m=+1114.166821601" observedRunningTime="2025-11-23 07:08:22.100309775 +0000 UTC m=+1117.613701078" watchObservedRunningTime="2025-11-23 07:08:22.130404091 +0000 UTC m=+1117.643795394" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.222882 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 23 07:08:22 crc kubenswrapper[4906]: I1123 07:08:22.258571 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.065991 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ftw6h" event={"ID":"e34ddce9-d32e-4c31-9716-86ad39a590e9","Type":"ContainerStarted","Data":"45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b"} Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.067873 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.067909 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.100791 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-ftw6h" podStartSLOduration=26.875145598 podStartE2EDuration="32.100758137s" podCreationTimestamp="2025-11-23 07:07:51 +0000 UTC" firstStartedPulling="2025-11-23 07:08:13.163383864 +0000 UTC m=+1108.676775187" lastFinishedPulling="2025-11-23 07:08:18.388996423 +0000 UTC m=+1113.902387726" observedRunningTime="2025-11-23 07:08:23.097616734 +0000 UTC m=+1118.611008037" watchObservedRunningTime="2025-11-23 07:08:23.100758137 +0000 UTC m=+1118.614149430" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.123868 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.125543 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.398533 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.400034 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.404290 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.404451 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.404572 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.404894 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-fpp7d" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.424145 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.428198 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-scripts\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.428298 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.428320 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.428336 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.428493 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-config\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.428602 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc5rs\" (UniqueName: \"kubernetes.io/projected/63999fb5-88e7-4df1-8084-267d0e37ac4c-kube-api-access-fc5rs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.428882 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.531306 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.531359 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.531384 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.531415 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-config\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.531449 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc5rs\" (UniqueName: \"kubernetes.io/projected/63999fb5-88e7-4df1-8084-267d0e37ac4c-kube-api-access-fc5rs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.531499 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.531564 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-scripts\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.532505 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.532632 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-scripts\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.532788 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-config\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.539221 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.540191 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.546342 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.557776 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc5rs\" (UniqueName: \"kubernetes.io/projected/63999fb5-88e7-4df1-8084-267d0e37ac4c-kube-api-access-fc5rs\") pod \"ovn-northd-0\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " pod="openstack/ovn-northd-0" Nov 23 07:08:23 crc kubenswrapper[4906]: I1123 07:08:23.717801 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 07:08:24 crc kubenswrapper[4906]: I1123 07:08:24.075461 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:08:24 crc kubenswrapper[4906]: I1123 07:08:24.076084 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:08:24 crc kubenswrapper[4906]: I1123 07:08:24.163460 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 07:08:24 crc kubenswrapper[4906]: W1123 07:08:24.166855 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63999fb5_88e7_4df1_8084_267d0e37ac4c.slice/crio-86dd091db9ec1326251a0c8111f47d74f5b0e80c0c63ef7f27d6f3deb87e779f WatchSource:0}: Error finding container 86dd091db9ec1326251a0c8111f47d74f5b0e80c0c63ef7f27d6f3deb87e779f: Status 404 returned error can't find the container with id 86dd091db9ec1326251a0c8111f47d74f5b0e80c0c63ef7f27d6f3deb87e779f Nov 23 07:08:24 crc kubenswrapper[4906]: I1123 07:08:24.175264 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 23 07:08:24 crc kubenswrapper[4906]: I1123 07:08:24.175315 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 23 07:08:24 crc kubenswrapper[4906]: I1123 07:08:24.277315 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.102260 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63999fb5-88e7-4df1-8084-267d0e37ac4c","Type":"ContainerStarted","Data":"86dd091db9ec1326251a0c8111f47d74f5b0e80c0c63ef7f27d6f3deb87e779f"} Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.215574 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.263851 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.586260 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.586587 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.704268 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-g6275"] Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.705301 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g6275" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.712873 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b64f-account-create-vvmd6"] Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.714270 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.715983 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.717592 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.727642 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b64f-account-create-vvmd6"] Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.735938 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-g6275"] Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.775341 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6673033-6ef9-4a43-90ca-8e1312660b02-operator-scripts\") pod \"placement-db-create-g6275\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " pod="openstack/placement-db-create-g6275" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.775730 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c27q9\" (UniqueName: \"kubernetes.io/projected/b6673033-6ef9-4a43-90ca-8e1312660b02-kube-api-access-c27q9\") pod \"placement-db-create-g6275\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " pod="openstack/placement-db-create-g6275" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.877097 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6673033-6ef9-4a43-90ca-8e1312660b02-operator-scripts\") pod \"placement-db-create-g6275\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " pod="openstack/placement-db-create-g6275" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.877260 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c27q9\" (UniqueName: \"kubernetes.io/projected/b6673033-6ef9-4a43-90ca-8e1312660b02-kube-api-access-c27q9\") pod \"placement-db-create-g6275\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " pod="openstack/placement-db-create-g6275" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.877427 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4ca0afe-d33d-434b-9182-78ba0588c29e-operator-scripts\") pod \"placement-b64f-account-create-vvmd6\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.877551 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxwc4\" (UniqueName: \"kubernetes.io/projected/a4ca0afe-d33d-434b-9182-78ba0588c29e-kube-api-access-wxwc4\") pod \"placement-b64f-account-create-vvmd6\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.878288 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6673033-6ef9-4a43-90ca-8e1312660b02-operator-scripts\") pod \"placement-db-create-g6275\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " pod="openstack/placement-db-create-g6275" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.894835 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c27q9\" (UniqueName: \"kubernetes.io/projected/b6673033-6ef9-4a43-90ca-8e1312660b02-kube-api-access-c27q9\") pod \"placement-db-create-g6275\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " pod="openstack/placement-db-create-g6275" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.978766 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4ca0afe-d33d-434b-9182-78ba0588c29e-operator-scripts\") pod \"placement-b64f-account-create-vvmd6\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.978831 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxwc4\" (UniqueName: \"kubernetes.io/projected/a4ca0afe-d33d-434b-9182-78ba0588c29e-kube-api-access-wxwc4\") pod \"placement-b64f-account-create-vvmd6\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.982371 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4ca0afe-d33d-434b-9182-78ba0588c29e-operator-scripts\") pod \"placement-b64f-account-create-vvmd6\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:25 crc kubenswrapper[4906]: I1123 07:08:25.995532 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxwc4\" (UniqueName: \"kubernetes.io/projected/a4ca0afe-d33d-434b-9182-78ba0588c29e-kube-api-access-wxwc4\") pod \"placement-b64f-account-create-vvmd6\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.079845 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g6275" Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.085870 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.112665 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63999fb5-88e7-4df1-8084-267d0e37ac4c","Type":"ContainerStarted","Data":"631998df3fcc19a80bacb62ef5030d483398f5a45ac22c6b6964329c20351f4a"} Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.112720 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63999fb5-88e7-4df1-8084-267d0e37ac4c","Type":"ContainerStarted","Data":"e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65"} Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.137764 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.8375301830000002 podStartE2EDuration="3.137741462s" podCreationTimestamp="2025-11-23 07:08:23 +0000 UTC" firstStartedPulling="2025-11-23 07:08:24.170365133 +0000 UTC m=+1119.683756436" lastFinishedPulling="2025-11-23 07:08:25.470576412 +0000 UTC m=+1120.983967715" observedRunningTime="2025-11-23 07:08:26.135954204 +0000 UTC m=+1121.649345507" watchObservedRunningTime="2025-11-23 07:08:26.137741462 +0000 UTC m=+1121.651132755" Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.228111 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.560250 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b64f-account-create-vvmd6"] Nov 23 07:08:26 crc kubenswrapper[4906]: W1123 07:08:26.566004 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4ca0afe_d33d_434b_9182_78ba0588c29e.slice/crio-33b398e78c0c5b3295eef996b8419b7a86754f9df557d8a9cf36e8ba3ff84c59 WatchSource:0}: Error finding container 33b398e78c0c5b3295eef996b8419b7a86754f9df557d8a9cf36e8ba3ff84c59: Status 404 returned error can't find the container with id 33b398e78c0c5b3295eef996b8419b7a86754f9df557d8a9cf36e8ba3ff84c59 Nov 23 07:08:26 crc kubenswrapper[4906]: I1123 07:08:26.572888 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-g6275"] Nov 23 07:08:26 crc kubenswrapper[4906]: W1123 07:08:26.581951 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6673033_6ef9_4a43_90ca_8e1312660b02.slice/crio-a26ccf879a748bf3374803d7d18b715eccf5593641e43fe351ac6db156400e59 WatchSource:0}: Error finding container a26ccf879a748bf3374803d7d18b715eccf5593641e43fe351ac6db156400e59: Status 404 returned error can't find the container with id a26ccf879a748bf3374803d7d18b715eccf5593641e43fe351ac6db156400e59 Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.125254 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g6275" event={"ID":"b6673033-6ef9-4a43-90ca-8e1312660b02","Type":"ContainerDied","Data":"8f173df3c1376722f31ab2e49d813520a0aa789fb2f6fe19a15bde7d99bd2b3b"} Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.125729 4906 generic.go:334] "Generic (PLEG): container finished" podID="b6673033-6ef9-4a43-90ca-8e1312660b02" containerID="8f173df3c1376722f31ab2e49d813520a0aa789fb2f6fe19a15bde7d99bd2b3b" exitCode=0 Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.125778 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g6275" event={"ID":"b6673033-6ef9-4a43-90ca-8e1312660b02","Type":"ContainerStarted","Data":"a26ccf879a748bf3374803d7d18b715eccf5593641e43fe351ac6db156400e59"} Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.128568 4906 generic.go:334] "Generic (PLEG): container finished" podID="a4ca0afe-d33d-434b-9182-78ba0588c29e" containerID="9e4e6012c2cef7cf10c2d0a9aa750a115c7cac5349940c88ef4bec142dc282da" exitCode=0 Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.128669 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b64f-account-create-vvmd6" event={"ID":"a4ca0afe-d33d-434b-9182-78ba0588c29e","Type":"ContainerDied","Data":"9e4e6012c2cef7cf10c2d0a9aa750a115c7cac5349940c88ef4bec142dc282da"} Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.128763 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b64f-account-create-vvmd6" event={"ID":"a4ca0afe-d33d-434b-9182-78ba0588c29e","Type":"ContainerStarted","Data":"33b398e78c0c5b3295eef996b8419b7a86754f9df557d8a9cf36e8ba3ff84c59"} Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.129493 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.354730 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-hdlmk"] Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.355162 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" podUID="38c90719-7ad4-4864-ab04-fe94caecd409" containerName="dnsmasq-dns" containerID="cri-o://dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16" gracePeriod=10 Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.356382 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.393851 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-b6tpv"] Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.395342 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.423192 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-b6tpv"] Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.510909 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r44b7\" (UniqueName: \"kubernetes.io/projected/0cd9d13e-9b11-44a7-9605-5e6d85825993-kube-api-access-r44b7\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.510974 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.511028 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-config\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.511110 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.511149 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.612843 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r44b7\" (UniqueName: \"kubernetes.io/projected/0cd9d13e-9b11-44a7-9605-5e6d85825993-kube-api-access-r44b7\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.612910 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.612968 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-config\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.613055 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.613111 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.614557 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.614591 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-config\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.614633 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.615235 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.636726 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r44b7\" (UniqueName: \"kubernetes.io/projected/0cd9d13e-9b11-44a7-9605-5e6d85825993-kube-api-access-r44b7\") pod \"dnsmasq-dns-5c9fdb784c-b6tpv\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.761074 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.860344 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.919522 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpk5b\" (UniqueName: \"kubernetes.io/projected/38c90719-7ad4-4864-ab04-fe94caecd409-kube-api-access-cpk5b\") pod \"38c90719-7ad4-4864-ab04-fe94caecd409\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.919709 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-sb\") pod \"38c90719-7ad4-4864-ab04-fe94caecd409\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.919762 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-dns-svc\") pod \"38c90719-7ad4-4864-ab04-fe94caecd409\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.919813 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-config\") pod \"38c90719-7ad4-4864-ab04-fe94caecd409\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.919840 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-nb\") pod \"38c90719-7ad4-4864-ab04-fe94caecd409\" (UID: \"38c90719-7ad4-4864-ab04-fe94caecd409\") " Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.960425 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c90719-7ad4-4864-ab04-fe94caecd409-kube-api-access-cpk5b" (OuterVolumeSpecName: "kube-api-access-cpk5b") pod "38c90719-7ad4-4864-ab04-fe94caecd409" (UID: "38c90719-7ad4-4864-ab04-fe94caecd409"). InnerVolumeSpecName "kube-api-access-cpk5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.987636 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-config" (OuterVolumeSpecName: "config") pod "38c90719-7ad4-4864-ab04-fe94caecd409" (UID: "38c90719-7ad4-4864-ab04-fe94caecd409"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.989060 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "38c90719-7ad4-4864-ab04-fe94caecd409" (UID: "38c90719-7ad4-4864-ab04-fe94caecd409"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:27 crc kubenswrapper[4906]: I1123 07:08:27.989850 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "38c90719-7ad4-4864-ab04-fe94caecd409" (UID: "38c90719-7ad4-4864-ab04-fe94caecd409"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.022570 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpk5b\" (UniqueName: \"kubernetes.io/projected/38c90719-7ad4-4864-ab04-fe94caecd409-kube-api-access-cpk5b\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.022614 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.022625 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.022634 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.025101 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "38c90719-7ad4-4864-ab04-fe94caecd409" (UID: "38c90719-7ad4-4864-ab04-fe94caecd409"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.123899 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38c90719-7ad4-4864-ab04-fe94caecd409-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.155157 4906 generic.go:334] "Generic (PLEG): container finished" podID="38c90719-7ad4-4864-ab04-fe94caecd409" containerID="dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16" exitCode=0 Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.155339 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" event={"ID":"38c90719-7ad4-4864-ab04-fe94caecd409","Type":"ContainerDied","Data":"dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16"} Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.155424 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" event={"ID":"38c90719-7ad4-4864-ab04-fe94caecd409","Type":"ContainerDied","Data":"472ba886c332b94bf313b14c820fadb76102ef49ba60fa0d3f58862847fb3f3d"} Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.155446 4906 scope.go:117] "RemoveContainer" containerID="dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.155588 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-hdlmk" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.193330 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-hdlmk"] Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.205370 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-hdlmk"] Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.206881 4906 scope.go:117] "RemoveContainer" containerID="c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.232932 4906 scope.go:117] "RemoveContainer" containerID="dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16" Nov 23 07:08:28 crc kubenswrapper[4906]: E1123 07:08:28.233512 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16\": container with ID starting with dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16 not found: ID does not exist" containerID="dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.233551 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16"} err="failed to get container status \"dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16\": rpc error: code = NotFound desc = could not find container \"dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16\": container with ID starting with dff1c04aa89bb9246d283520fbea0a4d57a419fe846918383bd830303fdd6b16 not found: ID does not exist" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.233578 4906 scope.go:117] "RemoveContainer" containerID="c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625" Nov 23 07:08:28 crc kubenswrapper[4906]: E1123 07:08:28.233995 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625\": container with ID starting with c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625 not found: ID does not exist" containerID="c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.234022 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625"} err="failed to get container status \"c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625\": rpc error: code = NotFound desc = could not find container \"c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625\": container with ID starting with c1f63287f55f89cf01806dc1ff8c687559a1907db6c9afc0ae9b43c71a797625 not found: ID does not exist" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.341499 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-b6tpv"] Nov 23 07:08:28 crc kubenswrapper[4906]: W1123 07:08:28.353478 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cd9d13e_9b11_44a7_9605_5e6d85825993.slice/crio-1c380d7549e14ce1101d48d59433d0e8dc07b0e687a4598393eb74c8475a2b7e WatchSource:0}: Error finding container 1c380d7549e14ce1101d48d59433d0e8dc07b0e687a4598393eb74c8475a2b7e: Status 404 returned error can't find the container with id 1c380d7549e14ce1101d48d59433d0e8dc07b0e687a4598393eb74c8475a2b7e Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.519593 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 23 07:08:28 crc kubenswrapper[4906]: E1123 07:08:28.520395 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c90719-7ad4-4864-ab04-fe94caecd409" containerName="init" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.520415 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c90719-7ad4-4864-ab04-fe94caecd409" containerName="init" Nov 23 07:08:28 crc kubenswrapper[4906]: E1123 07:08:28.520448 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c90719-7ad4-4864-ab04-fe94caecd409" containerName="dnsmasq-dns" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.520455 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c90719-7ad4-4864-ab04-fe94caecd409" containerName="dnsmasq-dns" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.520631 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c90719-7ad4-4864-ab04-fe94caecd409" containerName="dnsmasq-dns" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.525959 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.530520 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-7xqkx" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.530582 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.530533 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.530912 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.563402 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.643001 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-cache\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.643112 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.643172 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-lock\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.643674 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.643756 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lsl4\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-kube-api-access-7lsl4\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.669514 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g6275" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.672601 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.745552 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxwc4\" (UniqueName: \"kubernetes.io/projected/a4ca0afe-d33d-434b-9182-78ba0588c29e-kube-api-access-wxwc4\") pod \"a4ca0afe-d33d-434b-9182-78ba0588c29e\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.745693 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6673033-6ef9-4a43-90ca-8e1312660b02-operator-scripts\") pod \"b6673033-6ef9-4a43-90ca-8e1312660b02\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.745794 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c27q9\" (UniqueName: \"kubernetes.io/projected/b6673033-6ef9-4a43-90ca-8e1312660b02-kube-api-access-c27q9\") pod \"b6673033-6ef9-4a43-90ca-8e1312660b02\" (UID: \"b6673033-6ef9-4a43-90ca-8e1312660b02\") " Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.745971 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4ca0afe-d33d-434b-9182-78ba0588c29e-operator-scripts\") pod \"a4ca0afe-d33d-434b-9182-78ba0588c29e\" (UID: \"a4ca0afe-d33d-434b-9182-78ba0588c29e\") " Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746271 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lsl4\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-kube-api-access-7lsl4\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746352 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-cache\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746394 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746457 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4ca0afe-d33d-434b-9182-78ba0588c29e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a4ca0afe-d33d-434b-9182-78ba0588c29e" (UID: "a4ca0afe-d33d-434b-9182-78ba0588c29e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746521 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-lock\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746571 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746614 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4ca0afe-d33d-434b-9182-78ba0588c29e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: E1123 07:08:28.746799 4906 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 07:08:28 crc kubenswrapper[4906]: E1123 07:08:28.746818 4906 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 07:08:28 crc kubenswrapper[4906]: E1123 07:08:28.746867 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift podName:c4ec7c62-702e-4794-81fc-64c72f72cb12 nodeName:}" failed. No retries permitted until 2025-11-23 07:08:29.246852475 +0000 UTC m=+1124.760243778 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift") pod "swift-storage-0" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12") : configmap "swift-ring-files" not found Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746861 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-cache\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.746896 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.747087 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6673033-6ef9-4a43-90ca-8e1312660b02-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6673033-6ef9-4a43-90ca-8e1312660b02" (UID: "b6673033-6ef9-4a43-90ca-8e1312660b02"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.747328 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-lock\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.752063 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6673033-6ef9-4a43-90ca-8e1312660b02-kube-api-access-c27q9" (OuterVolumeSpecName: "kube-api-access-c27q9") pod "b6673033-6ef9-4a43-90ca-8e1312660b02" (UID: "b6673033-6ef9-4a43-90ca-8e1312660b02"). InnerVolumeSpecName "kube-api-access-c27q9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.752379 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ca0afe-d33d-434b-9182-78ba0588c29e-kube-api-access-wxwc4" (OuterVolumeSpecName: "kube-api-access-wxwc4") pod "a4ca0afe-d33d-434b-9182-78ba0588c29e" (UID: "a4ca0afe-d33d-434b-9182-78ba0588c29e"). InnerVolumeSpecName "kube-api-access-wxwc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.765022 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lsl4\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-kube-api-access-7lsl4\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.767577 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.848467 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6673033-6ef9-4a43-90ca-8e1312660b02-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.848498 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c27q9\" (UniqueName: \"kubernetes.io/projected/b6673033-6ef9-4a43-90ca-8e1312660b02-kube-api-access-c27q9\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:28 crc kubenswrapper[4906]: I1123 07:08:28.848513 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxwc4\" (UniqueName: \"kubernetes.io/projected/a4ca0afe-d33d-434b-9182-78ba0588c29e-kube-api-access-wxwc4\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.165594 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g6275" event={"ID":"b6673033-6ef9-4a43-90ca-8e1312660b02","Type":"ContainerDied","Data":"a26ccf879a748bf3374803d7d18b715eccf5593641e43fe351ac6db156400e59"} Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.165643 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a26ccf879a748bf3374803d7d18b715eccf5593641e43fe351ac6db156400e59" Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.165726 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g6275" Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.178125 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e368b3d0-8ac9-49be-b2c4-8e88d8327811","Type":"ContainerStarted","Data":"3d34a0cfbf6cce02048082a86c433bac46bdbc467dfe235995d608ebba0146d0"} Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.179087 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.182965 4906 generic.go:334] "Generic (PLEG): container finished" podID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerID="10f7a5e65693fad4378780c52506ab5c80d765de6888636c2b598b1ec596f52c" exitCode=0 Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.183049 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" event={"ID":"0cd9d13e-9b11-44a7-9605-5e6d85825993","Type":"ContainerDied","Data":"10f7a5e65693fad4378780c52506ab5c80d765de6888636c2b598b1ec596f52c"} Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.183077 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" event={"ID":"0cd9d13e-9b11-44a7-9605-5e6d85825993","Type":"ContainerStarted","Data":"1c380d7549e14ce1101d48d59433d0e8dc07b0e687a4598393eb74c8475a2b7e"} Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.185345 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b64f-account-create-vvmd6" event={"ID":"a4ca0afe-d33d-434b-9182-78ba0588c29e","Type":"ContainerDied","Data":"33b398e78c0c5b3295eef996b8419b7a86754f9df557d8a9cf36e8ba3ff84c59"} Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.185373 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33b398e78c0c5b3295eef996b8419b7a86754f9df557d8a9cf36e8ba3ff84c59" Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.185416 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b64f-account-create-vvmd6" Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.198344 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.55684968 podStartE2EDuration="42.198329613s" podCreationTimestamp="2025-11-23 07:07:47 +0000 UTC" firstStartedPulling="2025-11-23 07:07:48.127137591 +0000 UTC m=+1083.640528894" lastFinishedPulling="2025-11-23 07:08:28.768617534 +0000 UTC m=+1124.282008827" observedRunningTime="2025-11-23 07:08:29.195286743 +0000 UTC m=+1124.708678076" watchObservedRunningTime="2025-11-23 07:08:29.198329613 +0000 UTC m=+1124.711720916" Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.257327 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:29 crc kubenswrapper[4906]: E1123 07:08:29.257513 4906 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 07:08:29 crc kubenswrapper[4906]: E1123 07:08:29.257548 4906 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 07:08:29 crc kubenswrapper[4906]: E1123 07:08:29.257616 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift podName:c4ec7c62-702e-4794-81fc-64c72f72cb12 nodeName:}" failed. No retries permitted until 2025-11-23 07:08:30.257595381 +0000 UTC m=+1125.770986684 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift") pod "swift-storage-0" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12") : configmap "swift-ring-files" not found Nov 23 07:08:29 crc kubenswrapper[4906]: I1123 07:08:29.382312 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c90719-7ad4-4864-ab04-fe94caecd409" path="/var/lib/kubelet/pods/38c90719-7ad4-4864-ab04-fe94caecd409/volumes" Nov 23 07:08:30 crc kubenswrapper[4906]: I1123 07:08:30.194299 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" event={"ID":"0cd9d13e-9b11-44a7-9605-5e6d85825993","Type":"ContainerStarted","Data":"c22abeb905abc2533734bfb50f54b7f8670728d975ded0e99476ac3685b3c6dc"} Nov 23 07:08:30 crc kubenswrapper[4906]: I1123 07:08:30.194793 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:30 crc kubenswrapper[4906]: I1123 07:08:30.217409 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" podStartSLOduration=3.21738078 podStartE2EDuration="3.21738078s" podCreationTimestamp="2025-11-23 07:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:08:30.210511661 +0000 UTC m=+1125.723902964" watchObservedRunningTime="2025-11-23 07:08:30.21738078 +0000 UTC m=+1125.730772083" Nov 23 07:08:30 crc kubenswrapper[4906]: I1123 07:08:30.275865 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:30 crc kubenswrapper[4906]: E1123 07:08:30.276015 4906 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 07:08:30 crc kubenswrapper[4906]: E1123 07:08:30.276047 4906 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 07:08:30 crc kubenswrapper[4906]: E1123 07:08:30.276111 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift podName:c4ec7c62-702e-4794-81fc-64c72f72cb12 nodeName:}" failed. No retries permitted until 2025-11-23 07:08:32.276089263 +0000 UTC m=+1127.789480666 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift") pod "swift-storage-0" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12") : configmap "swift-ring-files" not found Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.008411 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-4srzx"] Nov 23 07:08:31 crc kubenswrapper[4906]: E1123 07:08:31.008987 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ca0afe-d33d-434b-9182-78ba0588c29e" containerName="mariadb-account-create" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.009015 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ca0afe-d33d-434b-9182-78ba0588c29e" containerName="mariadb-account-create" Nov 23 07:08:31 crc kubenswrapper[4906]: E1123 07:08:31.009033 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6673033-6ef9-4a43-90ca-8e1312660b02" containerName="mariadb-database-create" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.009044 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6673033-6ef9-4a43-90ca-8e1312660b02" containerName="mariadb-database-create" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.009284 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ca0afe-d33d-434b-9182-78ba0588c29e" containerName="mariadb-account-create" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.009312 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6673033-6ef9-4a43-90ca-8e1312660b02" containerName="mariadb-database-create" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.010134 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.022144 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4srzx"] Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.091787 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996f74a-36a3-4c18-a588-ff251e5b40fc-operator-scripts\") pod \"glance-db-create-4srzx\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.091858 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw952\" (UniqueName: \"kubernetes.io/projected/a996f74a-36a3-4c18-a588-ff251e5b40fc-kube-api-access-xw952\") pod \"glance-db-create-4srzx\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.119032 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-aaf0-account-create-8drwt"] Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.120435 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.123216 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.129345 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-aaf0-account-create-8drwt"] Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.193446 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pngmx\" (UniqueName: \"kubernetes.io/projected/2510029e-afba-42bb-b00d-8aac204d6b22-kube-api-access-pngmx\") pod \"glance-aaf0-account-create-8drwt\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.193600 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2510029e-afba-42bb-b00d-8aac204d6b22-operator-scripts\") pod \"glance-aaf0-account-create-8drwt\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.194224 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996f74a-36a3-4c18-a588-ff251e5b40fc-operator-scripts\") pod \"glance-db-create-4srzx\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.194270 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw952\" (UniqueName: \"kubernetes.io/projected/a996f74a-36a3-4c18-a588-ff251e5b40fc-kube-api-access-xw952\") pod \"glance-db-create-4srzx\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.195263 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996f74a-36a3-4c18-a588-ff251e5b40fc-operator-scripts\") pod \"glance-db-create-4srzx\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.211795 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw952\" (UniqueName: \"kubernetes.io/projected/a996f74a-36a3-4c18-a588-ff251e5b40fc-kube-api-access-xw952\") pod \"glance-db-create-4srzx\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.296243 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pngmx\" (UniqueName: \"kubernetes.io/projected/2510029e-afba-42bb-b00d-8aac204d6b22-kube-api-access-pngmx\") pod \"glance-aaf0-account-create-8drwt\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.296890 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2510029e-afba-42bb-b00d-8aac204d6b22-operator-scripts\") pod \"glance-aaf0-account-create-8drwt\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.297844 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2510029e-afba-42bb-b00d-8aac204d6b22-operator-scripts\") pod \"glance-aaf0-account-create-8drwt\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.315300 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pngmx\" (UniqueName: \"kubernetes.io/projected/2510029e-afba-42bb-b00d-8aac204d6b22-kube-api-access-pngmx\") pod \"glance-aaf0-account-create-8drwt\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.331273 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4srzx" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.436617 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.787088 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4srzx"] Nov 23 07:08:31 crc kubenswrapper[4906]: W1123 07:08:31.792360 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda996f74a_36a3_4c18_a588_ff251e5b40fc.slice/crio-429e507776ce4b81623fc58fffeed23b0c16dd8e79f3f86afc55e0c70499e32c WatchSource:0}: Error finding container 429e507776ce4b81623fc58fffeed23b0c16dd8e79f3f86afc55e0c70499e32c: Status 404 returned error can't find the container with id 429e507776ce4b81623fc58fffeed23b0c16dd8e79f3f86afc55e0c70499e32c Nov 23 07:08:31 crc kubenswrapper[4906]: I1123 07:08:31.915453 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-aaf0-account-create-8drwt"] Nov 23 07:08:31 crc kubenswrapper[4906]: W1123 07:08:31.925166 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2510029e_afba_42bb_b00d_8aac204d6b22.slice/crio-353983cc0f03915351e20b6a767bc7e67d1e54568db6e3e85fe1d936e730befe WatchSource:0}: Error finding container 353983cc0f03915351e20b6a767bc7e67d1e54568db6e3e85fe1d936e730befe: Status 404 returned error can't find the container with id 353983cc0f03915351e20b6a767bc7e67d1e54568db6e3e85fe1d936e730befe Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.220392 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aaf0-account-create-8drwt" event={"ID":"2510029e-afba-42bb-b00d-8aac204d6b22","Type":"ContainerStarted","Data":"717d5d2bacd9b96545d2a48c905f3b13c01a02ade9d241f21d8207a70788d9a9"} Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.220444 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aaf0-account-create-8drwt" event={"ID":"2510029e-afba-42bb-b00d-8aac204d6b22","Type":"ContainerStarted","Data":"353983cc0f03915351e20b6a767bc7e67d1e54568db6e3e85fe1d936e730befe"} Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.223134 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4srzx" event={"ID":"a996f74a-36a3-4c18-a588-ff251e5b40fc","Type":"ContainerStarted","Data":"89459c3df93e572c8aa54d6cdf68fb56123719eb60818d2b1a1c63833b8221fa"} Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.223189 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4srzx" event={"ID":"a996f74a-36a3-4c18-a588-ff251e5b40fc","Type":"ContainerStarted","Data":"429e507776ce4b81623fc58fffeed23b0c16dd8e79f3f86afc55e0c70499e32c"} Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.238018 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-aaf0-account-create-8drwt" podStartSLOduration=1.238002018 podStartE2EDuration="1.238002018s" podCreationTimestamp="2025-11-23 07:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:08:32.23653774 +0000 UTC m=+1127.749929063" watchObservedRunningTime="2025-11-23 07:08:32.238002018 +0000 UTC m=+1127.751393331" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.260539 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-4srzx" podStartSLOduration=2.260515557 podStartE2EDuration="2.260515557s" podCreationTimestamp="2025-11-23 07:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:08:32.257282201 +0000 UTC m=+1127.770673504" watchObservedRunningTime="2025-11-23 07:08:32.260515557 +0000 UTC m=+1127.773906860" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.336665 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:32 crc kubenswrapper[4906]: E1123 07:08:32.336979 4906 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 07:08:32 crc kubenswrapper[4906]: E1123 07:08:32.337245 4906 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 07:08:32 crc kubenswrapper[4906]: E1123 07:08:32.337393 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift podName:c4ec7c62-702e-4794-81fc-64c72f72cb12 nodeName:}" failed. No retries permitted until 2025-11-23 07:08:36.337335091 +0000 UTC m=+1131.850726454 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift") pod "swift-storage-0" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12") : configmap "swift-ring-files" not found Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.486672 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-tcrg7"] Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.487661 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.489754 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.489885 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.490473 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.505626 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tcrg7"] Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.541183 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35bc1bb0-6a22-4beb-bd74-111e73116437-etc-swift\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.541280 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cfv7\" (UniqueName: \"kubernetes.io/projected/35bc1bb0-6a22-4beb-bd74-111e73116437-kube-api-access-6cfv7\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.541361 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-scripts\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.541397 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-ring-data-devices\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.541457 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-combined-ca-bundle\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.541529 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-swiftconf\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.541557 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-dispersionconf\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.642917 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-ring-data-devices\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.643004 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-combined-ca-bundle\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.643071 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-swiftconf\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.643099 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-dispersionconf\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.643159 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35bc1bb0-6a22-4beb-bd74-111e73116437-etc-swift\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.643192 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cfv7\" (UniqueName: \"kubernetes.io/projected/35bc1bb0-6a22-4beb-bd74-111e73116437-kube-api-access-6cfv7\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.643229 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-scripts\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.644137 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-scripts\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.644188 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35bc1bb0-6a22-4beb-bd74-111e73116437-etc-swift\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.644740 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-ring-data-devices\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.657616 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-swiftconf\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.661775 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-combined-ca-bundle\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.677387 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-dispersionconf\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.697471 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cfv7\" (UniqueName: \"kubernetes.io/projected/35bc1bb0-6a22-4beb-bd74-111e73116437-kube-api-access-6cfv7\") pod \"swift-ring-rebalance-tcrg7\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:32 crc kubenswrapper[4906]: I1123 07:08:32.805410 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:33 crc kubenswrapper[4906]: I1123 07:08:33.240190 4906 generic.go:334] "Generic (PLEG): container finished" podID="2510029e-afba-42bb-b00d-8aac204d6b22" containerID="717d5d2bacd9b96545d2a48c905f3b13c01a02ade9d241f21d8207a70788d9a9" exitCode=0 Nov 23 07:08:33 crc kubenswrapper[4906]: I1123 07:08:33.240341 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aaf0-account-create-8drwt" event={"ID":"2510029e-afba-42bb-b00d-8aac204d6b22","Type":"ContainerDied","Data":"717d5d2bacd9b96545d2a48c905f3b13c01a02ade9d241f21d8207a70788d9a9"} Nov 23 07:08:33 crc kubenswrapper[4906]: I1123 07:08:33.241797 4906 generic.go:334] "Generic (PLEG): container finished" podID="a996f74a-36a3-4c18-a588-ff251e5b40fc" containerID="89459c3df93e572c8aa54d6cdf68fb56123719eb60818d2b1a1c63833b8221fa" exitCode=0 Nov 23 07:08:33 crc kubenswrapper[4906]: I1123 07:08:33.241821 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4srzx" event={"ID":"a996f74a-36a3-4c18-a588-ff251e5b40fc","Type":"ContainerDied","Data":"89459c3df93e572c8aa54d6cdf68fb56123719eb60818d2b1a1c63833b8221fa"} Nov 23 07:08:33 crc kubenswrapper[4906]: W1123 07:08:33.282878 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35bc1bb0_6a22_4beb_bd74_111e73116437.slice/crio-9538b99fabb57ed922e07fd0114fa714d399803e3089ed2fa477eb89273e3197 WatchSource:0}: Error finding container 9538b99fabb57ed922e07fd0114fa714d399803e3089ed2fa477eb89273e3197: Status 404 returned error can't find the container with id 9538b99fabb57ed922e07fd0114fa714d399803e3089ed2fa477eb89273e3197 Nov 23 07:08:33 crc kubenswrapper[4906]: I1123 07:08:33.296381 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tcrg7"] Nov 23 07:08:34 crc kubenswrapper[4906]: I1123 07:08:34.255759 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcrg7" event={"ID":"35bc1bb0-6a22-4beb-bd74-111e73116437","Type":"ContainerStarted","Data":"9538b99fabb57ed922e07fd0114fa714d399803e3089ed2fa477eb89273e3197"} Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.274036 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-f9bjf"] Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.275408 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.284254 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-f9bjf"] Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.402699 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61603290-df20-4ffb-af90-16aef4437224-operator-scripts\") pod \"keystone-db-create-f9bjf\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.402878 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnvfl\" (UniqueName: \"kubernetes.io/projected/61603290-df20-4ffb-af90-16aef4437224-kube-api-access-wnvfl\") pod \"keystone-db-create-f9bjf\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.423602 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-92c3-account-create-25d6n"] Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.424958 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.430732 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.444357 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-92c3-account-create-25d6n"] Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.504636 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjgrn\" (UniqueName: \"kubernetes.io/projected/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-kube-api-access-cjgrn\") pod \"keystone-92c3-account-create-25d6n\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.504889 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61603290-df20-4ffb-af90-16aef4437224-operator-scripts\") pod \"keystone-db-create-f9bjf\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.505036 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-operator-scripts\") pod \"keystone-92c3-account-create-25d6n\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.505156 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnvfl\" (UniqueName: \"kubernetes.io/projected/61603290-df20-4ffb-af90-16aef4437224-kube-api-access-wnvfl\") pod \"keystone-db-create-f9bjf\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.506306 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61603290-df20-4ffb-af90-16aef4437224-operator-scripts\") pod \"keystone-db-create-f9bjf\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.526183 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnvfl\" (UniqueName: \"kubernetes.io/projected/61603290-df20-4ffb-af90-16aef4437224-kube-api-access-wnvfl\") pod \"keystone-db-create-f9bjf\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.607023 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjgrn\" (UniqueName: \"kubernetes.io/projected/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-kube-api-access-cjgrn\") pod \"keystone-92c3-account-create-25d6n\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.607146 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-operator-scripts\") pod \"keystone-92c3-account-create-25d6n\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.608030 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-operator-scripts\") pod \"keystone-92c3-account-create-25d6n\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.611378 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.632462 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjgrn\" (UniqueName: \"kubernetes.io/projected/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-kube-api-access-cjgrn\") pod \"keystone-92c3-account-create-25d6n\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:35 crc kubenswrapper[4906]: I1123 07:08:35.755138 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.286194 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4srzx" event={"ID":"a996f74a-36a3-4c18-a588-ff251e5b40fc","Type":"ContainerDied","Data":"429e507776ce4b81623fc58fffeed23b0c16dd8e79f3f86afc55e0c70499e32c"} Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.286603 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="429e507776ce4b81623fc58fffeed23b0c16dd8e79f3f86afc55e0c70499e32c" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.288452 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.289452 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aaf0-account-create-8drwt" event={"ID":"2510029e-afba-42bb-b00d-8aac204d6b22","Type":"ContainerDied","Data":"353983cc0f03915351e20b6a767bc7e67d1e54568db6e3e85fe1d936e730befe"} Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.289473 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="353983cc0f03915351e20b6a767bc7e67d1e54568db6e3e85fe1d936e730befe" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.293053 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4srzx" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.419934 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2510029e-afba-42bb-b00d-8aac204d6b22-operator-scripts\") pod \"2510029e-afba-42bb-b00d-8aac204d6b22\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.420106 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw952\" (UniqueName: \"kubernetes.io/projected/a996f74a-36a3-4c18-a588-ff251e5b40fc-kube-api-access-xw952\") pod \"a996f74a-36a3-4c18-a588-ff251e5b40fc\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.420197 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pngmx\" (UniqueName: \"kubernetes.io/projected/2510029e-afba-42bb-b00d-8aac204d6b22-kube-api-access-pngmx\") pod \"2510029e-afba-42bb-b00d-8aac204d6b22\" (UID: \"2510029e-afba-42bb-b00d-8aac204d6b22\") " Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.420246 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996f74a-36a3-4c18-a588-ff251e5b40fc-operator-scripts\") pod \"a996f74a-36a3-4c18-a588-ff251e5b40fc\" (UID: \"a996f74a-36a3-4c18-a588-ff251e5b40fc\") " Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.420580 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:36 crc kubenswrapper[4906]: E1123 07:08:36.420972 4906 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 07:08:36 crc kubenswrapper[4906]: E1123 07:08:36.421018 4906 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 07:08:36 crc kubenswrapper[4906]: E1123 07:08:36.421125 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift podName:c4ec7c62-702e-4794-81fc-64c72f72cb12 nodeName:}" failed. No retries permitted until 2025-11-23 07:08:44.421107508 +0000 UTC m=+1139.934498811 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift") pod "swift-storage-0" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12") : configmap "swift-ring-files" not found Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.421195 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2510029e-afba-42bb-b00d-8aac204d6b22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2510029e-afba-42bb-b00d-8aac204d6b22" (UID: "2510029e-afba-42bb-b00d-8aac204d6b22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.421195 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a996f74a-36a3-4c18-a588-ff251e5b40fc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a996f74a-36a3-4c18-a588-ff251e5b40fc" (UID: "a996f74a-36a3-4c18-a588-ff251e5b40fc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.426628 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a996f74a-36a3-4c18-a588-ff251e5b40fc-kube-api-access-xw952" (OuterVolumeSpecName: "kube-api-access-xw952") pod "a996f74a-36a3-4c18-a588-ff251e5b40fc" (UID: "a996f74a-36a3-4c18-a588-ff251e5b40fc"). InnerVolumeSpecName "kube-api-access-xw952". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.444292 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2510029e-afba-42bb-b00d-8aac204d6b22-kube-api-access-pngmx" (OuterVolumeSpecName: "kube-api-access-pngmx") pod "2510029e-afba-42bb-b00d-8aac204d6b22" (UID: "2510029e-afba-42bb-b00d-8aac204d6b22"). InnerVolumeSpecName "kube-api-access-pngmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.522801 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a996f74a-36a3-4c18-a588-ff251e5b40fc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.522833 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2510029e-afba-42bb-b00d-8aac204d6b22-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.522845 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw952\" (UniqueName: \"kubernetes.io/projected/a996f74a-36a3-4c18-a588-ff251e5b40fc-kube-api-access-xw952\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:36 crc kubenswrapper[4906]: I1123 07:08:36.522856 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pngmx\" (UniqueName: \"kubernetes.io/projected/2510029e-afba-42bb-b00d-8aac204d6b22-kube-api-access-pngmx\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:37 crc kubenswrapper[4906]: I1123 07:08:37.296837 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aaf0-account-create-8drwt" Nov 23 07:08:37 crc kubenswrapper[4906]: I1123 07:08:37.298789 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4srzx" Nov 23 07:08:37 crc kubenswrapper[4906]: I1123 07:08:37.421911 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 23 07:08:37 crc kubenswrapper[4906]: I1123 07:08:37.763944 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:08:37 crc kubenswrapper[4906]: I1123 07:08:37.870922 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-xgt22"] Nov 23 07:08:37 crc kubenswrapper[4906]: I1123 07:08:37.871583 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" podUID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerName="dnsmasq-dns" containerID="cri-o://34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44" gracePeriod=10 Nov 23 07:08:38 crc kubenswrapper[4906]: I1123 07:08:38.777955 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.296716 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.347885 4906 generic.go:334] "Generic (PLEG): container finished" podID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerID="34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44" exitCode=0 Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.348057 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.348156 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" event={"ID":"607acb77-3bfa-4e7c-8aef-07597f2f2e1a","Type":"ContainerDied","Data":"34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44"} Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.348224 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-xgt22" event={"ID":"607acb77-3bfa-4e7c-8aef-07597f2f2e1a","Type":"ContainerDied","Data":"1bca03efe4961b4a867c44cefc65931ca7576266356308eb273b4411ba7ef572"} Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.348248 4906 scope.go:117] "RemoveContainer" containerID="34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.378086 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcrg7" event={"ID":"35bc1bb0-6a22-4beb-bd74-111e73116437","Type":"ContainerStarted","Data":"adb454b88595fdda4ef098a37bd75a106172393c0263a8fabed2363a86dda08c"} Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.388237 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlgsg\" (UniqueName: \"kubernetes.io/projected/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-kube-api-access-zlgsg\") pod \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.388374 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-ovsdbserver-sb\") pod \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.388402 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-dns-svc\") pod \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.388453 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-config\") pod \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\" (UID: \"607acb77-3bfa-4e7c-8aef-07597f2f2e1a\") " Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.402230 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-tcrg7" podStartSLOduration=1.656413772 podStartE2EDuration="7.402209754s" podCreationTimestamp="2025-11-23 07:08:32 +0000 UTC" firstStartedPulling="2025-11-23 07:08:33.284445171 +0000 UTC m=+1128.797836474" lastFinishedPulling="2025-11-23 07:08:39.030241153 +0000 UTC m=+1134.543632456" observedRunningTime="2025-11-23 07:08:39.393108887 +0000 UTC m=+1134.906500190" watchObservedRunningTime="2025-11-23 07:08:39.402209754 +0000 UTC m=+1134.915601057" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.414543 4906 scope.go:117] "RemoveContainer" containerID="b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.421409 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-kube-api-access-zlgsg" (OuterVolumeSpecName: "kube-api-access-zlgsg") pod "607acb77-3bfa-4e7c-8aef-07597f2f2e1a" (UID: "607acb77-3bfa-4e7c-8aef-07597f2f2e1a"). InnerVolumeSpecName "kube-api-access-zlgsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.484101 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "607acb77-3bfa-4e7c-8aef-07597f2f2e1a" (UID: "607acb77-3bfa-4e7c-8aef-07597f2f2e1a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.493709 4906 scope.go:117] "RemoveContainer" containerID="34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44" Nov 23 07:08:39 crc kubenswrapper[4906]: E1123 07:08:39.494281 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44\": container with ID starting with 34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44 not found: ID does not exist" containerID="34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.494374 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44"} err="failed to get container status \"34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44\": rpc error: code = NotFound desc = could not find container \"34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44\": container with ID starting with 34f60bac7444e53f82b914a1bd481908c168d1f0c52c5e3a84301059c0c92b44 not found: ID does not exist" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.494480 4906 scope.go:117] "RemoveContainer" containerID="b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.494425 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlgsg\" (UniqueName: \"kubernetes.io/projected/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-kube-api-access-zlgsg\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.504952 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:39 crc kubenswrapper[4906]: E1123 07:08:39.506650 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668\": container with ID starting with b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668 not found: ID does not exist" containerID="b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.506790 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668"} err="failed to get container status \"b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668\": rpc error: code = NotFound desc = could not find container \"b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668\": container with ID starting with b036790ec55839aa675015cece482ff96a688dde4bd555ceb05c27ce77774668 not found: ID does not exist" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.555206 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-config" (OuterVolumeSpecName: "config") pod "607acb77-3bfa-4e7c-8aef-07597f2f2e1a" (UID: "607acb77-3bfa-4e7c-8aef-07597f2f2e1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.571187 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-92c3-account-create-25d6n"] Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.573363 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "607acb77-3bfa-4e7c-8aef-07597f2f2e1a" (UID: "607acb77-3bfa-4e7c-8aef-07597f2f2e1a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.604768 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-f9bjf"] Nov 23 07:08:39 crc kubenswrapper[4906]: W1123 07:08:39.606309 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61603290_df20_4ffb_af90_16aef4437224.slice/crio-ddc7a707b27e221a3779787303e872fc031dbd813e81c7014d0619db18a9c16d WatchSource:0}: Error finding container ddc7a707b27e221a3779787303e872fc031dbd813e81c7014d0619db18a9c16d: Status 404 returned error can't find the container with id ddc7a707b27e221a3779787303e872fc031dbd813e81c7014d0619db18a9c16d Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.607016 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.607032 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/607acb77-3bfa-4e7c-8aef-07597f2f2e1a-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.709590 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-xgt22"] Nov 23 07:08:39 crc kubenswrapper[4906]: I1123 07:08:39.716407 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-xgt22"] Nov 23 07:08:40 crc kubenswrapper[4906]: I1123 07:08:40.367087 4906 generic.go:334] "Generic (PLEG): container finished" podID="55501b57-eb6b-4753-a9bd-43ff19f6c7a9" containerID="bf5e01bb5e89dd0c74104dbe793e0f8882ba08a29b84959fea106fea381a3521" exitCode=0 Nov 23 07:08:40 crc kubenswrapper[4906]: I1123 07:08:40.367308 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-92c3-account-create-25d6n" event={"ID":"55501b57-eb6b-4753-a9bd-43ff19f6c7a9","Type":"ContainerDied","Data":"bf5e01bb5e89dd0c74104dbe793e0f8882ba08a29b84959fea106fea381a3521"} Nov 23 07:08:40 crc kubenswrapper[4906]: I1123 07:08:40.367470 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-92c3-account-create-25d6n" event={"ID":"55501b57-eb6b-4753-a9bd-43ff19f6c7a9","Type":"ContainerStarted","Data":"ef751ee431435518ce78e057b21f9634d62073681be3953d7783a79cd3869bce"} Nov 23 07:08:40 crc kubenswrapper[4906]: I1123 07:08:40.369110 4906 generic.go:334] "Generic (PLEG): container finished" podID="61603290-df20-4ffb-af90-16aef4437224" containerID="3acf2c2a799d15e29b4ea27d19732e6c4972f398d8d2402a22dee98452de758a" exitCode=0 Nov 23 07:08:40 crc kubenswrapper[4906]: I1123 07:08:40.369191 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f9bjf" event={"ID":"61603290-df20-4ffb-af90-16aef4437224","Type":"ContainerDied","Data":"3acf2c2a799d15e29b4ea27d19732e6c4972f398d8d2402a22dee98452de758a"} Nov 23 07:08:40 crc kubenswrapper[4906]: I1123 07:08:40.369257 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f9bjf" event={"ID":"61603290-df20-4ffb-af90-16aef4437224","Type":"ContainerStarted","Data":"ddc7a707b27e221a3779787303e872fc031dbd813e81c7014d0619db18a9c16d"} Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.262466 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-d2j47"] Nov 23 07:08:41 crc kubenswrapper[4906]: E1123 07:08:41.262922 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerName="init" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.262951 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerName="init" Nov 23 07:08:41 crc kubenswrapper[4906]: E1123 07:08:41.262985 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerName="dnsmasq-dns" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.262996 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerName="dnsmasq-dns" Nov 23 07:08:41 crc kubenswrapper[4906]: E1123 07:08:41.263017 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a996f74a-36a3-4c18-a588-ff251e5b40fc" containerName="mariadb-database-create" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.263026 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a996f74a-36a3-4c18-a588-ff251e5b40fc" containerName="mariadb-database-create" Nov 23 07:08:41 crc kubenswrapper[4906]: E1123 07:08:41.263042 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2510029e-afba-42bb-b00d-8aac204d6b22" containerName="mariadb-account-create" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.263051 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2510029e-afba-42bb-b00d-8aac204d6b22" containerName="mariadb-account-create" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.263274 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2510029e-afba-42bb-b00d-8aac204d6b22" containerName="mariadb-account-create" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.263874 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" containerName="dnsmasq-dns" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.263899 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a996f74a-36a3-4c18-a588-ff251e5b40fc" containerName="mariadb-database-create" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.264596 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.267410 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5v68l" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.267776 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.274829 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-d2j47"] Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.336594 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-db-sync-config-data\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.336782 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkdpm\" (UniqueName: \"kubernetes.io/projected/32ddb74f-97de-4c04-bdca-efbcf93e5840-kube-api-access-bkdpm\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.336826 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-config-data\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.336851 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-combined-ca-bundle\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.367878 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="607acb77-3bfa-4e7c-8aef-07597f2f2e1a" path="/var/lib/kubelet/pods/607acb77-3bfa-4e7c-8aef-07597f2f2e1a/volumes" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.438555 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-db-sync-config-data\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.439038 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkdpm\" (UniqueName: \"kubernetes.io/projected/32ddb74f-97de-4c04-bdca-efbcf93e5840-kube-api-access-bkdpm\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.439077 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-config-data\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.439106 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-combined-ca-bundle\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.448556 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-db-sync-config-data\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.448569 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-config-data\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.460498 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-combined-ca-bundle\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.460830 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkdpm\" (UniqueName: \"kubernetes.io/projected/32ddb74f-97de-4c04-bdca-efbcf93e5840-kube-api-access-bkdpm\") pod \"glance-db-sync-d2j47\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.586637 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d2j47" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.765886 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.833881 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.854908 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjgrn\" (UniqueName: \"kubernetes.io/projected/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-kube-api-access-cjgrn\") pod \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.855061 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-operator-scripts\") pod \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\" (UID: \"55501b57-eb6b-4753-a9bd-43ff19f6c7a9\") " Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.855894 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55501b57-eb6b-4753-a9bd-43ff19f6c7a9" (UID: "55501b57-eb6b-4753-a9bd-43ff19f6c7a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.859093 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-kube-api-access-cjgrn" (OuterVolumeSpecName: "kube-api-access-cjgrn") pod "55501b57-eb6b-4753-a9bd-43ff19f6c7a9" (UID: "55501b57-eb6b-4753-a9bd-43ff19f6c7a9"). InnerVolumeSpecName "kube-api-access-cjgrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.956437 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61603290-df20-4ffb-af90-16aef4437224-operator-scripts\") pod \"61603290-df20-4ffb-af90-16aef4437224\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.956586 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnvfl\" (UniqueName: \"kubernetes.io/projected/61603290-df20-4ffb-af90-16aef4437224-kube-api-access-wnvfl\") pod \"61603290-df20-4ffb-af90-16aef4437224\" (UID: \"61603290-df20-4ffb-af90-16aef4437224\") " Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.957082 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.957105 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjgrn\" (UniqueName: \"kubernetes.io/projected/55501b57-eb6b-4753-a9bd-43ff19f6c7a9-kube-api-access-cjgrn\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.957389 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61603290-df20-4ffb-af90-16aef4437224-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "61603290-df20-4ffb-af90-16aef4437224" (UID: "61603290-df20-4ffb-af90-16aef4437224"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:41 crc kubenswrapper[4906]: I1123 07:08:41.960341 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61603290-df20-4ffb-af90-16aef4437224-kube-api-access-wnvfl" (OuterVolumeSpecName: "kube-api-access-wnvfl") pod "61603290-df20-4ffb-af90-16aef4437224" (UID: "61603290-df20-4ffb-af90-16aef4437224"). InnerVolumeSpecName "kube-api-access-wnvfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.064964 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnvfl\" (UniqueName: \"kubernetes.io/projected/61603290-df20-4ffb-af90-16aef4437224-kube-api-access-wnvfl\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.064993 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61603290-df20-4ffb-af90-16aef4437224-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.247414 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-d2j47"] Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.384388 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f9bjf" event={"ID":"61603290-df20-4ffb-af90-16aef4437224","Type":"ContainerDied","Data":"ddc7a707b27e221a3779787303e872fc031dbd813e81c7014d0619db18a9c16d"} Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.384742 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddc7a707b27e221a3779787303e872fc031dbd813e81c7014d0619db18a9c16d" Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.384433 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f9bjf" Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.385907 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-92c3-account-create-25d6n" event={"ID":"55501b57-eb6b-4753-a9bd-43ff19f6c7a9","Type":"ContainerDied","Data":"ef751ee431435518ce78e057b21f9634d62073681be3953d7783a79cd3869bce"} Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.385944 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef751ee431435518ce78e057b21f9634d62073681be3953d7783a79cd3869bce" Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.385981 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-92c3-account-create-25d6n" Nov 23 07:08:42 crc kubenswrapper[4906]: I1123 07:08:42.393106 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d2j47" event={"ID":"32ddb74f-97de-4c04-bdca-efbcf93e5840","Type":"ContainerStarted","Data":"8e0d65f169bc07592c3c98529697283e606b68cb8c9e041cd73fb1c5b06301ac"} Nov 23 07:08:44 crc kubenswrapper[4906]: I1123 07:08:44.509972 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:08:44 crc kubenswrapper[4906]: E1123 07:08:44.510224 4906 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 07:08:44 crc kubenswrapper[4906]: E1123 07:08:44.510468 4906 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 07:08:44 crc kubenswrapper[4906]: E1123 07:08:44.510527 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift podName:c4ec7c62-702e-4794-81fc-64c72f72cb12 nodeName:}" failed. No retries permitted until 2025-11-23 07:09:00.510510061 +0000 UTC m=+1156.023901364 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift") pod "swift-storage-0" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12") : configmap "swift-ring-files" not found Nov 23 07:08:47 crc kubenswrapper[4906]: I1123 07:08:47.430662 4906 generic.go:334] "Generic (PLEG): container finished" podID="35bc1bb0-6a22-4beb-bd74-111e73116437" containerID="adb454b88595fdda4ef098a37bd75a106172393c0263a8fabed2363a86dda08c" exitCode=0 Nov 23 07:08:47 crc kubenswrapper[4906]: I1123 07:08:47.430722 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcrg7" event={"ID":"35bc1bb0-6a22-4beb-bd74-111e73116437","Type":"ContainerDied","Data":"adb454b88595fdda4ef098a37bd75a106172393c0263a8fabed2363a86dda08c"} Nov 23 07:08:49 crc kubenswrapper[4906]: I1123 07:08:49.449032 4906 generic.go:334] "Generic (PLEG): container finished" podID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerID="e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44" exitCode=0 Nov 23 07:08:49 crc kubenswrapper[4906]: I1123 07:08:49.449382 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a0222ea-7767-4d08-a6ec-6659a33f9df2","Type":"ContainerDied","Data":"e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44"} Nov 23 07:08:51 crc kubenswrapper[4906]: I1123 07:08:51.798756 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l9d4x" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" probeResult="failure" output=< Nov 23 07:08:51 crc kubenswrapper[4906]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 23 07:08:51 crc kubenswrapper[4906]: > Nov 23 07:08:51 crc kubenswrapper[4906]: I1123 07:08:51.823626 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:08:52 crc kubenswrapper[4906]: I1123 07:08:52.488315 4906 generic.go:334] "Generic (PLEG): container finished" podID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerID="89f5735b86ee117a6dd64d5fbfd44ffb2a0e87dd340dc5bc635c3b876a04901a" exitCode=0 Nov 23 07:08:52 crc kubenswrapper[4906]: I1123 07:08:52.488788 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d677a57-1acf-4627-aa66-f4bee96a2b51","Type":"ContainerDied","Data":"89f5735b86ee117a6dd64d5fbfd44ffb2a0e87dd340dc5bc635c3b876a04901a"} Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.500194 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tcrg7" event={"ID":"35bc1bb0-6a22-4beb-bd74-111e73116437","Type":"ContainerDied","Data":"9538b99fabb57ed922e07fd0114fa714d399803e3089ed2fa477eb89273e3197"} Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.501017 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9538b99fabb57ed922e07fd0114fa714d399803e3089ed2fa477eb89273e3197" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.549393 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.678088 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35bc1bb0-6a22-4beb-bd74-111e73116437-etc-swift\") pod \"35bc1bb0-6a22-4beb-bd74-111e73116437\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.678248 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-scripts\") pod \"35bc1bb0-6a22-4beb-bd74-111e73116437\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.678358 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-swiftconf\") pod \"35bc1bb0-6a22-4beb-bd74-111e73116437\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.678402 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-ring-data-devices\") pod \"35bc1bb0-6a22-4beb-bd74-111e73116437\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.678465 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-dispersionconf\") pod \"35bc1bb0-6a22-4beb-bd74-111e73116437\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.678521 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-combined-ca-bundle\") pod \"35bc1bb0-6a22-4beb-bd74-111e73116437\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.678546 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cfv7\" (UniqueName: \"kubernetes.io/projected/35bc1bb0-6a22-4beb-bd74-111e73116437-kube-api-access-6cfv7\") pod \"35bc1bb0-6a22-4beb-bd74-111e73116437\" (UID: \"35bc1bb0-6a22-4beb-bd74-111e73116437\") " Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.680512 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "35bc1bb0-6a22-4beb-bd74-111e73116437" (UID: "35bc1bb0-6a22-4beb-bd74-111e73116437"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.681498 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35bc1bb0-6a22-4beb-bd74-111e73116437-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "35bc1bb0-6a22-4beb-bd74-111e73116437" (UID: "35bc1bb0-6a22-4beb-bd74-111e73116437"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.686804 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35bc1bb0-6a22-4beb-bd74-111e73116437-kube-api-access-6cfv7" (OuterVolumeSpecName: "kube-api-access-6cfv7") pod "35bc1bb0-6a22-4beb-bd74-111e73116437" (UID: "35bc1bb0-6a22-4beb-bd74-111e73116437"). InnerVolumeSpecName "kube-api-access-6cfv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.690786 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "35bc1bb0-6a22-4beb-bd74-111e73116437" (UID: "35bc1bb0-6a22-4beb-bd74-111e73116437"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.707380 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "35bc1bb0-6a22-4beb-bd74-111e73116437" (UID: "35bc1bb0-6a22-4beb-bd74-111e73116437"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.716350 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35bc1bb0-6a22-4beb-bd74-111e73116437" (UID: "35bc1bb0-6a22-4beb-bd74-111e73116437"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.716799 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-scripts" (OuterVolumeSpecName: "scripts") pod "35bc1bb0-6a22-4beb-bd74-111e73116437" (UID: "35bc1bb0-6a22-4beb-bd74-111e73116437"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.781372 4906 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.781415 4906 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.781430 4906 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.781442 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bc1bb0-6a22-4beb-bd74-111e73116437-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.781455 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cfv7\" (UniqueName: \"kubernetes.io/projected/35bc1bb0-6a22-4beb-bd74-111e73116437-kube-api-access-6cfv7\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.781468 4906 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/35bc1bb0-6a22-4beb-bd74-111e73116437-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:53 crc kubenswrapper[4906]: I1123 07:08:53.781478 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/35bc1bb0-6a22-4beb-bd74-111e73116437-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.512718 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d2j47" event={"ID":"32ddb74f-97de-4c04-bdca-efbcf93e5840","Type":"ContainerStarted","Data":"4defce0e3fb0237aa4ffc1b4b5e58328d1a52cca3c133f1db7f360d81d75d7ac"} Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.518427 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a0222ea-7767-4d08-a6ec-6659a33f9df2","Type":"ContainerStarted","Data":"bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af"} Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.518914 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.521217 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d677a57-1acf-4627-aa66-f4bee96a2b51","Type":"ContainerStarted","Data":"d83546ce8058eec4db9f5c5b6441b7ce247b7a874eb9b2018665ca353691af54"} Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.521247 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tcrg7" Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.547181 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-d2j47" podStartSLOduration=2.354428265 podStartE2EDuration="13.547160145s" podCreationTimestamp="2025-11-23 07:08:41 +0000 UTC" firstStartedPulling="2025-11-23 07:08:42.246719734 +0000 UTC m=+1137.760111037" lastFinishedPulling="2025-11-23 07:08:53.439451614 +0000 UTC m=+1148.952842917" observedRunningTime="2025-11-23 07:08:54.54158416 +0000 UTC m=+1150.054975473" watchObservedRunningTime="2025-11-23 07:08:54.547160145 +0000 UTC m=+1150.060551448" Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.569557 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371963.285236 podStartE2EDuration="1m13.56953869s" podCreationTimestamp="2025-11-23 07:07:41 +0000 UTC" firstStartedPulling="2025-11-23 07:07:43.174373474 +0000 UTC m=+1078.687764777" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:08:54.565610557 +0000 UTC m=+1150.079001860" watchObservedRunningTime="2025-11-23 07:08:54.56953869 +0000 UTC m=+1150.082929993" Nov 23 07:08:54 crc kubenswrapper[4906]: I1123 07:08:54.593173 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.952702451 podStartE2EDuration="1m14.593148316s" podCreationTimestamp="2025-11-23 07:07:40 +0000 UTC" firstStartedPulling="2025-11-23 07:07:42.320446531 +0000 UTC m=+1077.833837834" lastFinishedPulling="2025-11-23 07:08:14.960892396 +0000 UTC m=+1110.474283699" observedRunningTime="2025-11-23 07:08:54.588505255 +0000 UTC m=+1150.101896558" watchObservedRunningTime="2025-11-23 07:08:54.593148316 +0000 UTC m=+1150.106539619" Nov 23 07:08:56 crc kubenswrapper[4906]: I1123 07:08:56.808466 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l9d4x" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" probeResult="failure" output=< Nov 23 07:08:56 crc kubenswrapper[4906]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 23 07:08:56 crc kubenswrapper[4906]: > Nov 23 07:08:56 crc kubenswrapper[4906]: I1123 07:08:56.882846 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.221528 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l9d4x-config-92m4t"] Nov 23 07:08:57 crc kubenswrapper[4906]: E1123 07:08:57.222320 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61603290-df20-4ffb-af90-16aef4437224" containerName="mariadb-database-create" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.222340 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="61603290-df20-4ffb-af90-16aef4437224" containerName="mariadb-database-create" Nov 23 07:08:57 crc kubenswrapper[4906]: E1123 07:08:57.222381 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55501b57-eb6b-4753-a9bd-43ff19f6c7a9" containerName="mariadb-account-create" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.222387 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="55501b57-eb6b-4753-a9bd-43ff19f6c7a9" containerName="mariadb-account-create" Nov 23 07:08:57 crc kubenswrapper[4906]: E1123 07:08:57.222403 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35bc1bb0-6a22-4beb-bd74-111e73116437" containerName="swift-ring-rebalance" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.222411 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="35bc1bb0-6a22-4beb-bd74-111e73116437" containerName="swift-ring-rebalance" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.222572 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="35bc1bb0-6a22-4beb-bd74-111e73116437" containerName="swift-ring-rebalance" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.222585 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="61603290-df20-4ffb-af90-16aef4437224" containerName="mariadb-database-create" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.222599 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="55501b57-eb6b-4753-a9bd-43ff19f6c7a9" containerName="mariadb-account-create" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.223309 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.229860 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.248206 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l9d4x-config-92m4t"] Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.253349 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h84rp\" (UniqueName: \"kubernetes.io/projected/5826870b-c1aa-486d-9d9e-ff3203cf7c50-kube-api-access-h84rp\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.253608 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-log-ovn\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.253667 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-scripts\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.253726 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-additional-scripts\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.253767 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.253787 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run-ovn\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.355366 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-scripts\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.356728 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-additional-scripts\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.356821 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.356858 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run-ovn\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.356960 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h84rp\" (UniqueName: \"kubernetes.io/projected/5826870b-c1aa-486d-9d9e-ff3203cf7c50-kube-api-access-h84rp\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.357022 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-log-ovn\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.357497 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-log-ovn\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.357976 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run-ovn\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.358109 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.358545 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-additional-scripts\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.358549 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-scripts\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.402878 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h84rp\" (UniqueName: \"kubernetes.io/projected/5826870b-c1aa-486d-9d9e-ff3203cf7c50-kube-api-access-h84rp\") pod \"ovn-controller-l9d4x-config-92m4t\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:57 crc kubenswrapper[4906]: I1123 07:08:57.558803 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:08:58 crc kubenswrapper[4906]: I1123 07:08:58.042844 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l9d4x-config-92m4t"] Nov 23 07:08:58 crc kubenswrapper[4906]: W1123 07:08:58.052151 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5826870b_c1aa_486d_9d9e_ff3203cf7c50.slice/crio-8668b7a3f5000e505d0c5d7bac1f9d2f10e1fc8c1918aed39113f268ccd9a5ec WatchSource:0}: Error finding container 8668b7a3f5000e505d0c5d7bac1f9d2f10e1fc8c1918aed39113f268ccd9a5ec: Status 404 returned error can't find the container with id 8668b7a3f5000e505d0c5d7bac1f9d2f10e1fc8c1918aed39113f268ccd9a5ec Nov 23 07:08:58 crc kubenswrapper[4906]: I1123 07:08:58.559995 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x-config-92m4t" event={"ID":"5826870b-c1aa-486d-9d9e-ff3203cf7c50","Type":"ContainerStarted","Data":"eaab7475fcf8e8c5b70f5983da2df20065a5cb1d25322e150eb55c5ace653871"} Nov 23 07:08:58 crc kubenswrapper[4906]: I1123 07:08:58.560456 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x-config-92m4t" event={"ID":"5826870b-c1aa-486d-9d9e-ff3203cf7c50","Type":"ContainerStarted","Data":"8668b7a3f5000e505d0c5d7bac1f9d2f10e1fc8c1918aed39113f268ccd9a5ec"} Nov 23 07:08:59 crc kubenswrapper[4906]: I1123 07:08:59.573134 4906 generic.go:334] "Generic (PLEG): container finished" podID="5826870b-c1aa-486d-9d9e-ff3203cf7c50" containerID="eaab7475fcf8e8c5b70f5983da2df20065a5cb1d25322e150eb55c5ace653871" exitCode=0 Nov 23 07:08:59 crc kubenswrapper[4906]: I1123 07:08:59.573259 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x-config-92m4t" event={"ID":"5826870b-c1aa-486d-9d9e-ff3203cf7c50","Type":"ContainerDied","Data":"eaab7475fcf8e8c5b70f5983da2df20065a5cb1d25322e150eb55c5ace653871"} Nov 23 07:08:59 crc kubenswrapper[4906]: I1123 07:08:59.936426 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.028876 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h84rp\" (UniqueName: \"kubernetes.io/projected/5826870b-c1aa-486d-9d9e-ff3203cf7c50-kube-api-access-h84rp\") pod \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.028972 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run-ovn\") pod \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029069 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5826870b-c1aa-486d-9d9e-ff3203cf7c50" (UID: "5826870b-c1aa-486d-9d9e-ff3203cf7c50"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029147 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-scripts\") pod \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029273 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-log-ovn\") pod \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029298 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run\") pod \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029339 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5826870b-c1aa-486d-9d9e-ff3203cf7c50" (UID: "5826870b-c1aa-486d-9d9e-ff3203cf7c50"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029371 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-additional-scripts\") pod \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\" (UID: \"5826870b-c1aa-486d-9d9e-ff3203cf7c50\") " Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029407 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run" (OuterVolumeSpecName: "var-run") pod "5826870b-c1aa-486d-9d9e-ff3203cf7c50" (UID: "5826870b-c1aa-486d-9d9e-ff3203cf7c50"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029864 4906 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029878 4906 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.029886 4906 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5826870b-c1aa-486d-9d9e-ff3203cf7c50-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.030456 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-scripts" (OuterVolumeSpecName: "scripts") pod "5826870b-c1aa-486d-9d9e-ff3203cf7c50" (UID: "5826870b-c1aa-486d-9d9e-ff3203cf7c50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.030559 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5826870b-c1aa-486d-9d9e-ff3203cf7c50" (UID: "5826870b-c1aa-486d-9d9e-ff3203cf7c50"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.042761 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5826870b-c1aa-486d-9d9e-ff3203cf7c50-kube-api-access-h84rp" (OuterVolumeSpecName: "kube-api-access-h84rp") pod "5826870b-c1aa-486d-9d9e-ff3203cf7c50" (UID: "5826870b-c1aa-486d-9d9e-ff3203cf7c50"). InnerVolumeSpecName "kube-api-access-h84rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.131490 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.131534 4906 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5826870b-c1aa-486d-9d9e-ff3203cf7c50-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.131548 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h84rp\" (UniqueName: \"kubernetes.io/projected/5826870b-c1aa-486d-9d9e-ff3203cf7c50-kube-api-access-h84rp\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.539117 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.546211 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"swift-storage-0\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " pod="openstack/swift-storage-0" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.587656 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x-config-92m4t" event={"ID":"5826870b-c1aa-486d-9d9e-ff3203cf7c50","Type":"ContainerDied","Data":"8668b7a3f5000e505d0c5d7bac1f9d2f10e1fc8c1918aed39113f268ccd9a5ec"} Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.587852 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8668b7a3f5000e505d0c5d7bac1f9d2f10e1fc8c1918aed39113f268ccd9a5ec" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.587857 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x-config-92m4t" Nov 23 07:09:00 crc kubenswrapper[4906]: I1123 07:09:00.777984 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.081159 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l9d4x-config-92m4t"] Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.092694 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-l9d4x-config-92m4t"] Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.366324 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5826870b-c1aa-486d-9d9e-ff3203cf7c50" path="/var/lib/kubelet/pods/5826870b-c1aa-486d-9d9e-ff3203cf7c50/volumes" Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.367453 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 23 07:09:01 crc kubenswrapper[4906]: W1123 07:09:01.378482 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ec7c62_702e_4794_81fc_64c72f72cb12.slice/crio-90b0d500e80584b17478edcd992a6916fcff5fcf37c96acd17e254ed07ee3bd0 WatchSource:0}: Error finding container 90b0d500e80584b17478edcd992a6916fcff5fcf37c96acd17e254ed07ee3bd0: Status 404 returned error can't find the container with id 90b0d500e80584b17478edcd992a6916fcff5fcf37c96acd17e254ed07ee3bd0 Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.595849 4906 generic.go:334] "Generic (PLEG): container finished" podID="32ddb74f-97de-4c04-bdca-efbcf93e5840" containerID="4defce0e3fb0237aa4ffc1b4b5e58328d1a52cca3c133f1db7f360d81d75d7ac" exitCode=0 Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.595929 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d2j47" event={"ID":"32ddb74f-97de-4c04-bdca-efbcf93e5840","Type":"ContainerDied","Data":"4defce0e3fb0237aa4ffc1b4b5e58328d1a52cca3c133f1db7f360d81d75d7ac"} Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.597418 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"90b0d500e80584b17478edcd992a6916fcff5fcf37c96acd17e254ed07ee3bd0"} Nov 23 07:09:01 crc kubenswrapper[4906]: I1123 07:09:01.794304 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-l9d4x" Nov 23 07:09:02 crc kubenswrapper[4906]: I1123 07:09:02.688894 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.007107 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d2j47" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.097169 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-combined-ca-bundle\") pod \"32ddb74f-97de-4c04-bdca-efbcf93e5840\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.097625 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-db-sync-config-data\") pod \"32ddb74f-97de-4c04-bdca-efbcf93e5840\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.097860 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkdpm\" (UniqueName: \"kubernetes.io/projected/32ddb74f-97de-4c04-bdca-efbcf93e5840-kube-api-access-bkdpm\") pod \"32ddb74f-97de-4c04-bdca-efbcf93e5840\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.097918 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-config-data\") pod \"32ddb74f-97de-4c04-bdca-efbcf93e5840\" (UID: \"32ddb74f-97de-4c04-bdca-efbcf93e5840\") " Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.105522 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ddb74f-97de-4c04-bdca-efbcf93e5840-kube-api-access-bkdpm" (OuterVolumeSpecName: "kube-api-access-bkdpm") pod "32ddb74f-97de-4c04-bdca-efbcf93e5840" (UID: "32ddb74f-97de-4c04-bdca-efbcf93e5840"). InnerVolumeSpecName "kube-api-access-bkdpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.119975 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "32ddb74f-97de-4c04-bdca-efbcf93e5840" (UID: "32ddb74f-97de-4c04-bdca-efbcf93e5840"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.133073 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32ddb74f-97de-4c04-bdca-efbcf93e5840" (UID: "32ddb74f-97de-4c04-bdca-efbcf93e5840"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.161278 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-config-data" (OuterVolumeSpecName: "config-data") pod "32ddb74f-97de-4c04-bdca-efbcf93e5840" (UID: "32ddb74f-97de-4c04-bdca-efbcf93e5840"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.199978 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkdpm\" (UniqueName: \"kubernetes.io/projected/32ddb74f-97de-4c04-bdca-efbcf93e5840-kube-api-access-bkdpm\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.200015 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.200025 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.200034 4906 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ddb74f-97de-4c04-bdca-efbcf93e5840-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:03 crc kubenswrapper[4906]: E1123 07:09:03.482641 4906 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ddb74f_97de_4c04_bdca_efbcf93e5840.slice/crio-8e0d65f169bc07592c3c98529697283e606b68cb8c9e041cd73fb1c5b06301ac\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ddb74f_97de_4c04_bdca_efbcf93e5840.slice\": RecentStats: unable to find data in memory cache]" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.619413 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"11af2df38e7407624c492ea4718898da13ce0775329468c7c775d8b83ef7a458"} Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.619469 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"29fc0d262044ca645fb4133e01cbabc9655fe5b59a91b3028e0d08bcd5278256"} Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.619480 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"8bed0393fc8307e59d410b24ec754b21fe7d3bd1f3e83cd8d6f028d1a8f06cdd"} Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.619489 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"4f7149c3b0589b8222d668111ff70a5b576f33041ccbfe291e9f62085cc511d3"} Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.621068 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-d2j47" event={"ID":"32ddb74f-97de-4c04-bdca-efbcf93e5840","Type":"ContainerDied","Data":"8e0d65f169bc07592c3c98529697283e606b68cb8c9e041cd73fb1c5b06301ac"} Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.621097 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-d2j47" Nov 23 07:09:03 crc kubenswrapper[4906]: I1123 07:09:03.621111 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e0d65f169bc07592c3c98529697283e606b68cb8c9e041cd73fb1c5b06301ac" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.108819 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84ddf475bf-7v6q5"] Nov 23 07:09:04 crc kubenswrapper[4906]: E1123 07:09:04.109512 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5826870b-c1aa-486d-9d9e-ff3203cf7c50" containerName="ovn-config" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.109525 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5826870b-c1aa-486d-9d9e-ff3203cf7c50" containerName="ovn-config" Nov 23 07:09:04 crc kubenswrapper[4906]: E1123 07:09:04.109553 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ddb74f-97de-4c04-bdca-efbcf93e5840" containerName="glance-db-sync" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.109559 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ddb74f-97de-4c04-bdca-efbcf93e5840" containerName="glance-db-sync" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.109728 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5826870b-c1aa-486d-9d9e-ff3203cf7c50" containerName="ovn-config" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.109741 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ddb74f-97de-4c04-bdca-efbcf93e5840" containerName="glance-db-sync" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.110579 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.124191 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84ddf475bf-7v6q5"] Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.221312 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-sb\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.221439 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-config\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.221510 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-nb\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.221548 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk26f\" (UniqueName: \"kubernetes.io/projected/8bca5db5-3d68-45c7-88f6-8c933732d836-kube-api-access-qk26f\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.221975 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-dns-svc\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.323586 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-config\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.323656 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-nb\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.323699 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk26f\" (UniqueName: \"kubernetes.io/projected/8bca5db5-3d68-45c7-88f6-8c933732d836-kube-api-access-qk26f\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.323780 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-dns-svc\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.323806 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-sb\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.324766 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-config\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.324774 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-nb\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.324845 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-sb\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.325053 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-dns-svc\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.343661 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk26f\" (UniqueName: \"kubernetes.io/projected/8bca5db5-3d68-45c7-88f6-8c933732d836-kube-api-access-qk26f\") pod \"dnsmasq-dns-84ddf475bf-7v6q5\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:04 crc kubenswrapper[4906]: I1123 07:09:04.427328 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:05 crc kubenswrapper[4906]: I1123 07:09:05.026789 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84ddf475bf-7v6q5"] Nov 23 07:09:05 crc kubenswrapper[4906]: I1123 07:09:05.644324 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"0c11306e0c7841c07b12efafe09a24ad91a89cee793dba18a9f2876f0d47db2f"} Nov 23 07:09:05 crc kubenswrapper[4906]: I1123 07:09:05.647647 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" event={"ID":"8bca5db5-3d68-45c7-88f6-8c933732d836","Type":"ContainerDied","Data":"837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af"} Nov 23 07:09:05 crc kubenswrapper[4906]: I1123 07:09:05.647155 4906 generic.go:334] "Generic (PLEG): container finished" podID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerID="837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af" exitCode=0 Nov 23 07:09:05 crc kubenswrapper[4906]: I1123 07:09:05.647963 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" event={"ID":"8bca5db5-3d68-45c7-88f6-8c933732d836","Type":"ContainerStarted","Data":"b9c0b7599cbdd307d8d8924b2d9182986691dd6b5719adfc10ef32483efe1160"} Nov 23 07:09:06 crc kubenswrapper[4906]: I1123 07:09:06.659476 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"627197269c9d973c5e754357b5bea2cd6202f4ed8b342e5be634b4ffe2fc24f4"} Nov 23 07:09:06 crc kubenswrapper[4906]: I1123 07:09:06.659973 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"59f81af64ef288c720feec11a22a60062d378a09a432fe305317b0c5b6c0c0a3"} Nov 23 07:09:06 crc kubenswrapper[4906]: I1123 07:09:06.659984 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"d7765fe2025af755e1f8644c577f06f2002c8f768fc520d7b9ae7dfe318bd069"} Nov 23 07:09:06 crc kubenswrapper[4906]: I1123 07:09:06.663635 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" event={"ID":"8bca5db5-3d68-45c7-88f6-8c933732d836","Type":"ContainerStarted","Data":"2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9"} Nov 23 07:09:06 crc kubenswrapper[4906]: I1123 07:09:06.664235 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:06 crc kubenswrapper[4906]: I1123 07:09:06.711099 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" podStartSLOduration=2.711070883 podStartE2EDuration="2.711070883s" podCreationTimestamp="2025-11-23 07:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:06.704973073 +0000 UTC m=+1162.218364386" watchObservedRunningTime="2025-11-23 07:09:06.711070883 +0000 UTC m=+1162.224462186" Nov 23 07:09:08 crc kubenswrapper[4906]: I1123 07:09:08.710173 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"18a0cbb3bc39086918add992d8b7e36417c2a512ce34ad58af2473769333d276"} Nov 23 07:09:08 crc kubenswrapper[4906]: I1123 07:09:08.711009 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"7310698561fe415b603d4b23e1ca0b97cb10e0a5ab1b9e756910372da0037e1f"} Nov 23 07:09:08 crc kubenswrapper[4906]: I1123 07:09:08.711030 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"52c1da52f5e6b3969985d8979a7d057a2abc04e85e14b228a48cb576460bae01"} Nov 23 07:09:08 crc kubenswrapper[4906]: I1123 07:09:08.711045 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"28356b85be5f4f2918de3e09fcd6083deb2b49f2e911a40c85312e0eebf206a8"} Nov 23 07:09:08 crc kubenswrapper[4906]: I1123 07:09:08.711062 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"14e047c6504017d7ca6ee4d097e4dc0c1a3d8b79305ee2e36580fe0c8238c99b"} Nov 23 07:09:09 crc kubenswrapper[4906]: I1123 07:09:09.728818 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"0f3631e076f67c96972eae5e4520d40c2b4e9bafcc51ca7093e8a282b2d1ab2d"} Nov 23 07:09:09 crc kubenswrapper[4906]: I1123 07:09:09.729299 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerStarted","Data":"c1437bfa849a403a91c6260f5cd5aec30c660352ad87626d59e56d5664d49e22"} Nov 23 07:09:09 crc kubenswrapper[4906]: I1123 07:09:09.776977 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.554046862999996 podStartE2EDuration="42.776941861s" podCreationTimestamp="2025-11-23 07:08:27 +0000 UTC" firstStartedPulling="2025-11-23 07:09:01.382250288 +0000 UTC m=+1156.895641591" lastFinishedPulling="2025-11-23 07:09:07.605145276 +0000 UTC m=+1163.118536589" observedRunningTime="2025-11-23 07:09:09.769172849 +0000 UTC m=+1165.282564222" watchObservedRunningTime="2025-11-23 07:09:09.776941861 +0000 UTC m=+1165.290333204" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.086064 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84ddf475bf-7v6q5"] Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.086351 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" podUID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerName="dnsmasq-dns" containerID="cri-o://2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9" gracePeriod=10 Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.142401 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-hjdbc"] Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.145457 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.147877 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.167399 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-hjdbc"] Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.337009 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-sb\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.337110 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-swift-storage-0\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.337159 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-svc\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.337203 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-nb\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.337251 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45fz2\" (UniqueName: \"kubernetes.io/projected/d0e1a933-a98c-4401-99cd-5c1ad2873911-kube-api-access-45fz2\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.337278 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-config\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.440174 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-nb\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.440256 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45fz2\" (UniqueName: \"kubernetes.io/projected/d0e1a933-a98c-4401-99cd-5c1ad2873911-kube-api-access-45fz2\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.440292 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-config\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.440331 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-sb\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.440377 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-swift-storage-0\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.440412 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-svc\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.442106 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-svc\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.443052 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-nb\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.444508 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-config\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.445106 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-sb\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.447607 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-swift-storage-0\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.479082 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45fz2\" (UniqueName: \"kubernetes.io/projected/d0e1a933-a98c-4401-99cd-5c1ad2873911-kube-api-access-45fz2\") pod \"dnsmasq-dns-6856c564b9-hjdbc\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.496307 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.633304 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.658894 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk26f\" (UniqueName: \"kubernetes.io/projected/8bca5db5-3d68-45c7-88f6-8c933732d836-kube-api-access-qk26f\") pod \"8bca5db5-3d68-45c7-88f6-8c933732d836\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.658973 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-nb\") pod \"8bca5db5-3d68-45c7-88f6-8c933732d836\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.659047 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-dns-svc\") pod \"8bca5db5-3d68-45c7-88f6-8c933732d836\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.659158 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-sb\") pod \"8bca5db5-3d68-45c7-88f6-8c933732d836\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.659212 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-config\") pod \"8bca5db5-3d68-45c7-88f6-8c933732d836\" (UID: \"8bca5db5-3d68-45c7-88f6-8c933732d836\") " Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.729660 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bca5db5-3d68-45c7-88f6-8c933732d836-kube-api-access-qk26f" (OuterVolumeSpecName: "kube-api-access-qk26f") pod "8bca5db5-3d68-45c7-88f6-8c933732d836" (UID: "8bca5db5-3d68-45c7-88f6-8c933732d836"). InnerVolumeSpecName "kube-api-access-qk26f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.751259 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8bca5db5-3d68-45c7-88f6-8c933732d836" (UID: "8bca5db5-3d68-45c7-88f6-8c933732d836"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.752606 4906 generic.go:334] "Generic (PLEG): container finished" podID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerID="2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9" exitCode=0 Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.752804 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" event={"ID":"8bca5db5-3d68-45c7-88f6-8c933732d836","Type":"ContainerDied","Data":"2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9"} Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.752930 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" event={"ID":"8bca5db5-3d68-45c7-88f6-8c933732d836","Type":"ContainerDied","Data":"b9c0b7599cbdd307d8d8924b2d9182986691dd6b5719adfc10ef32483efe1160"} Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.752954 4906 scope.go:117] "RemoveContainer" containerID="2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.753715 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84ddf475bf-7v6q5" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.760851 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.760879 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk26f\" (UniqueName: \"kubernetes.io/projected/8bca5db5-3d68-45c7-88f6-8c933732d836-kube-api-access-qk26f\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.771004 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8bca5db5-3d68-45c7-88f6-8c933732d836" (UID: "8bca5db5-3d68-45c7-88f6-8c933732d836"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.771959 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-config" (OuterVolumeSpecName: "config") pod "8bca5db5-3d68-45c7-88f6-8c933732d836" (UID: "8bca5db5-3d68-45c7-88f6-8c933732d836"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.788636 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8bca5db5-3d68-45c7-88f6-8c933732d836" (UID: "8bca5db5-3d68-45c7-88f6-8c933732d836"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.791316 4906 scope.go:117] "RemoveContainer" containerID="837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.815983 4906 scope.go:117] "RemoveContainer" containerID="2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9" Nov 23 07:09:10 crc kubenswrapper[4906]: E1123 07:09:10.817968 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9\": container with ID starting with 2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9 not found: ID does not exist" containerID="2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.818005 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9"} err="failed to get container status \"2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9\": rpc error: code = NotFound desc = could not find container \"2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9\": container with ID starting with 2158cc7b1b21ad53a5bf7c5c75b7dff8c9f5056b0283892aee72447c2898ccd9 not found: ID does not exist" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.818036 4906 scope.go:117] "RemoveContainer" containerID="837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af" Nov 23 07:09:10 crc kubenswrapper[4906]: E1123 07:09:10.822944 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af\": container with ID starting with 837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af not found: ID does not exist" containerID="837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.823002 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af"} err="failed to get container status \"837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af\": rpc error: code = NotFound desc = could not find container \"837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af\": container with ID starting with 837052b1d33ac418aa9e00ea71b6c394d107d2d7960b66aa7a36f9170f52a4af not found: ID does not exist" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.864915 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.864951 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:10 crc kubenswrapper[4906]: I1123 07:09:10.864961 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bca5db5-3d68-45c7-88f6-8c933732d836-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:11 crc kubenswrapper[4906]: I1123 07:09:11.082183 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84ddf475bf-7v6q5"] Nov 23 07:09:11 crc kubenswrapper[4906]: I1123 07:09:11.088324 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84ddf475bf-7v6q5"] Nov 23 07:09:11 crc kubenswrapper[4906]: I1123 07:09:11.124000 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-hjdbc"] Nov 23 07:09:11 crc kubenswrapper[4906]: W1123 07:09:11.132022 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0e1a933_a98c_4401_99cd_5c1ad2873911.slice/crio-0d2a0502aefd5cedbddfdf97893e5e72684697537c09fda9a9ef69d5adf71c62 WatchSource:0}: Error finding container 0d2a0502aefd5cedbddfdf97893e5e72684697537c09fda9a9ef69d5adf71c62: Status 404 returned error can't find the container with id 0d2a0502aefd5cedbddfdf97893e5e72684697537c09fda9a9ef69d5adf71c62 Nov 23 07:09:11 crc kubenswrapper[4906]: I1123 07:09:11.369813 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bca5db5-3d68-45c7-88f6-8c933732d836" path="/var/lib/kubelet/pods/8bca5db5-3d68-45c7-88f6-8c933732d836/volumes" Nov 23 07:09:11 crc kubenswrapper[4906]: I1123 07:09:11.763446 4906 generic.go:334] "Generic (PLEG): container finished" podID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerID="2aa7e0410d9580fdc0f37c3e6733c159984772ff90d8d592f658e9ad0a53886b" exitCode=0 Nov 23 07:09:11 crc kubenswrapper[4906]: I1123 07:09:11.763521 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" event={"ID":"d0e1a933-a98c-4401-99cd-5c1ad2873911","Type":"ContainerDied","Data":"2aa7e0410d9580fdc0f37c3e6733c159984772ff90d8d592f658e9ad0a53886b"} Nov 23 07:09:11 crc kubenswrapper[4906]: I1123 07:09:11.763552 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" event={"ID":"d0e1a933-a98c-4401-99cd-5c1ad2873911","Type":"ContainerStarted","Data":"0d2a0502aefd5cedbddfdf97893e5e72684697537c09fda9a9ef69d5adf71c62"} Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.073956 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.399980 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-slkpb"] Nov 23 07:09:12 crc kubenswrapper[4906]: E1123 07:09:12.400472 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerName="init" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.400496 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerName="init" Nov 23 07:09:12 crc kubenswrapper[4906]: E1123 07:09:12.400514 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerName="dnsmasq-dns" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.400521 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerName="dnsmasq-dns" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.400781 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bca5db5-3d68-45c7-88f6-8c933732d836" containerName="dnsmasq-dns" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.407253 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.416166 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-slkpb"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.488905 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-vzfd7"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.489974 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.505185 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b28db04-7998-4e4c-80c2-dc2e246f55c5-operator-scripts\") pod \"cinder-db-create-slkpb\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.505290 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtskw\" (UniqueName: \"kubernetes.io/projected/0b28db04-7998-4e4c-80c2-dc2e246f55c5-kube-api-access-xtskw\") pod \"cinder-db-create-slkpb\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.517728 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vzfd7"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.537867 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a450-account-create-8nn5h"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.539468 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.545518 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.559353 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a450-account-create-8nn5h"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.607190 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9fa61a-edd3-476d-8084-10653351773b-operator-scripts\") pod \"barbican-db-create-vzfd7\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.607279 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b28db04-7998-4e4c-80c2-dc2e246f55c5-operator-scripts\") pod \"cinder-db-create-slkpb\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.607319 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtskw\" (UniqueName: \"kubernetes.io/projected/0b28db04-7998-4e4c-80c2-dc2e246f55c5-kube-api-access-xtskw\") pod \"cinder-db-create-slkpb\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.607345 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5w6p\" (UniqueName: \"kubernetes.io/projected/7e9fa61a-edd3-476d-8084-10653351773b-kube-api-access-l5w6p\") pod \"barbican-db-create-vzfd7\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.608132 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b28db04-7998-4e4c-80c2-dc2e246f55c5-operator-scripts\") pod \"cinder-db-create-slkpb\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.619642 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-46ad-account-create-wc7t2"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.620885 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.624403 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.626701 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-46ad-account-create-wc7t2"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.635375 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtskw\" (UniqueName: \"kubernetes.io/projected/0b28db04-7998-4e4c-80c2-dc2e246f55c5-kube-api-access-xtskw\") pod \"cinder-db-create-slkpb\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.693951 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.709351 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5w6p\" (UniqueName: \"kubernetes.io/projected/7e9fa61a-edd3-476d-8084-10653351773b-kube-api-access-l5w6p\") pod \"barbican-db-create-vzfd7\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.709449 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-operator-scripts\") pod \"barbican-a450-account-create-8nn5h\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.709523 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tkbs\" (UniqueName: \"kubernetes.io/projected/0c44516c-f839-4d45-b83d-50a0104d06d9-kube-api-access-4tkbs\") pod \"cinder-46ad-account-create-wc7t2\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.709601 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9fa61a-edd3-476d-8084-10653351773b-operator-scripts\") pod \"barbican-db-create-vzfd7\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.709653 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f2xl\" (UniqueName: \"kubernetes.io/projected/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-kube-api-access-8f2xl\") pod \"barbican-a450-account-create-8nn5h\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.709724 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c44516c-f839-4d45-b83d-50a0104d06d9-operator-scripts\") pod \"cinder-46ad-account-create-wc7t2\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.710387 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9fa61a-edd3-476d-8084-10653351773b-operator-scripts\") pod \"barbican-db-create-vzfd7\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.731304 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5w6p\" (UniqueName: \"kubernetes.io/projected/7e9fa61a-edd3-476d-8084-10653351773b-kube-api-access-l5w6p\") pod \"barbican-db-create-vzfd7\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.736263 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.831931 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.833803 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tkbs\" (UniqueName: \"kubernetes.io/projected/0c44516c-f839-4d45-b83d-50a0104d06d9-kube-api-access-4tkbs\") pod \"cinder-46ad-account-create-wc7t2\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.833907 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f2xl\" (UniqueName: \"kubernetes.io/projected/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-kube-api-access-8f2xl\") pod \"barbican-a450-account-create-8nn5h\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.833979 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c44516c-f839-4d45-b83d-50a0104d06d9-operator-scripts\") pod \"cinder-46ad-account-create-wc7t2\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.834052 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-operator-scripts\") pod \"barbican-a450-account-create-8nn5h\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.835043 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-operator-scripts\") pod \"barbican-a450-account-create-8nn5h\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.842456 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c44516c-f839-4d45-b83d-50a0104d06d9-operator-scripts\") pod \"cinder-46ad-account-create-wc7t2\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.856015 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" event={"ID":"d0e1a933-a98c-4401-99cd-5c1ad2873911","Type":"ContainerStarted","Data":"b4cc6a652fa7198326e843a437170ddeab2364ef675e70647d868d7c0942777a"} Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.857273 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.871868 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9f60-account-create-ljk4z"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.873968 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.901003 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.902037 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f2xl\" (UniqueName: \"kubernetes.io/projected/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-kube-api-access-8f2xl\") pod \"barbican-a450-account-create-8nn5h\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.907406 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9f60-account-create-ljk4z"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.937939 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-4vqvb"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.941557 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.966251 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" podStartSLOduration=2.966227493 podStartE2EDuration="2.966227493s" podCreationTimestamp="2025-11-23 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:12.951244351 +0000 UTC m=+1168.464635654" watchObservedRunningTime="2025-11-23 07:09:12.966227493 +0000 UTC m=+1168.479618796" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.967809 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tkbs\" (UniqueName: \"kubernetes.io/projected/0c44516c-f839-4d45-b83d-50a0104d06d9-kube-api-access-4tkbs\") pod \"cinder-46ad-account-create-wc7t2\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.987181 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4vqvb"] Nov 23 07:09:12 crc kubenswrapper[4906]: I1123 07:09:12.989980 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:12.998820 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-c7r8h"] Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.001883 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.017794 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.018878 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nk8ll" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.019010 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.019310 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.036667 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02f3febd-9b3c-4329-a767-16b438b44c91-operator-scripts\") pod \"neutron-9f60-account-create-ljk4z\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.036755 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nr69\" (UniqueName: \"kubernetes.io/projected/aac4be73-123c-4818-b96d-0a6ea8898fcb-kube-api-access-2nr69\") pod \"neutron-db-create-4vqvb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.036776 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac4be73-123c-4818-b96d-0a6ea8898fcb-operator-scripts\") pod \"neutron-db-create-4vqvb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.036862 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gggdj\" (UniqueName: \"kubernetes.io/projected/02f3febd-9b3c-4329-a767-16b438b44c91-kube-api-access-gggdj\") pod \"neutron-9f60-account-create-ljk4z\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.047362 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c7r8h"] Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.138969 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nr69\" (UniqueName: \"kubernetes.io/projected/aac4be73-123c-4818-b96d-0a6ea8898fcb-kube-api-access-2nr69\") pod \"neutron-db-create-4vqvb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.139019 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-combined-ca-bundle\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.139043 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac4be73-123c-4818-b96d-0a6ea8898fcb-operator-scripts\") pod \"neutron-db-create-4vqvb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.139112 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gggdj\" (UniqueName: \"kubernetes.io/projected/02f3febd-9b3c-4329-a767-16b438b44c91-kube-api-access-gggdj\") pod \"neutron-9f60-account-create-ljk4z\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.139142 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vn8c\" (UniqueName: \"kubernetes.io/projected/bb7e51ef-70f7-4fb2-a1ef-31a109125528-kube-api-access-9vn8c\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.139182 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-config-data\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.139214 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02f3febd-9b3c-4329-a767-16b438b44c91-operator-scripts\") pod \"neutron-9f60-account-create-ljk4z\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.140050 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02f3febd-9b3c-4329-a767-16b438b44c91-operator-scripts\") pod \"neutron-9f60-account-create-ljk4z\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.145202 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac4be73-123c-4818-b96d-0a6ea8898fcb-operator-scripts\") pod \"neutron-db-create-4vqvb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.160091 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.177546 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nr69\" (UniqueName: \"kubernetes.io/projected/aac4be73-123c-4818-b96d-0a6ea8898fcb-kube-api-access-2nr69\") pod \"neutron-db-create-4vqvb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.196312 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gggdj\" (UniqueName: \"kubernetes.io/projected/02f3febd-9b3c-4329-a767-16b438b44c91-kube-api-access-gggdj\") pod \"neutron-9f60-account-create-ljk4z\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.228280 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.243732 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vn8c\" (UniqueName: \"kubernetes.io/projected/bb7e51ef-70f7-4fb2-a1ef-31a109125528-kube-api-access-9vn8c\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.243823 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-config-data\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.243927 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-combined-ca-bundle\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.251657 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-combined-ca-bundle\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.258367 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-config-data\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.283264 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.291444 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vn8c\" (UniqueName: \"kubernetes.io/projected/bb7e51ef-70f7-4fb2-a1ef-31a109125528-kube-api-access-9vn8c\") pod \"keystone-db-sync-c7r8h\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.353309 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.590075 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-slkpb"] Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.803597 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vzfd7"] Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.866454 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vzfd7" event={"ID":"7e9fa61a-edd3-476d-8084-10653351773b","Type":"ContainerStarted","Data":"0f7357163617a64ee425d2b5481721b4b3e0a9c5f0739da3add0583400aaf2a0"} Nov 23 07:09:13 crc kubenswrapper[4906]: I1123 07:09:13.876268 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-slkpb" event={"ID":"0b28db04-7998-4e4c-80c2-dc2e246f55c5","Type":"ContainerStarted","Data":"4197fcea1f9b42c20cf7caa24964a54ff13d7bb534ed6f67d067d0a9f3547fe8"} Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.000343 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-46ad-account-create-wc7t2"] Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.130623 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a450-account-create-8nn5h"] Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.139803 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9f60-account-create-ljk4z"] Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.147802 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c7r8h"] Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.286275 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4vqvb"] Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.886770 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vzfd7" event={"ID":"7e9fa61a-edd3-476d-8084-10653351773b","Type":"ContainerStarted","Data":"018decda61999cccfd06ac4b20b5448abba52ae9229f9dd58c47095fdc6d2e57"} Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.888571 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f60-account-create-ljk4z" event={"ID":"02f3febd-9b3c-4329-a767-16b438b44c91","Type":"ContainerStarted","Data":"8e120d61e2177096d7bd7af35a9d0b9a7471d4cc69e23f1361986864ccca96c1"} Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.890584 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a450-account-create-8nn5h" event={"ID":"2c2f5a87-926b-4d48-8e44-ec8dc87f8462","Type":"ContainerStarted","Data":"3a7ba6fa7eaa4ccfc703b4704462f502740b8daaf5787cc5c410ff63a54492b7"} Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.892281 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4vqvb" event={"ID":"aac4be73-123c-4818-b96d-0a6ea8898fcb","Type":"ContainerStarted","Data":"9f2b4a5ca47c646e337d88bfac4adea25e7922173e247b5606a16f2311ffc372"} Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.893549 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c7r8h" event={"ID":"bb7e51ef-70f7-4fb2-a1ef-31a109125528","Type":"ContainerStarted","Data":"53e291ba23eae571244763f110de9045f9f507418c7aadbe815d57663611b1eb"} Nov 23 07:09:14 crc kubenswrapper[4906]: I1123 07:09:14.895588 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-46ad-account-create-wc7t2" event={"ID":"0c44516c-f839-4d45-b83d-50a0104d06d9","Type":"ContainerStarted","Data":"297e93b0f85689e56d1ced37a96a144993d3cd59153fafda192b6b95bed345cf"} Nov 23 07:09:16 crc kubenswrapper[4906]: I1123 07:09:16.920810 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-46ad-account-create-wc7t2" event={"ID":"0c44516c-f839-4d45-b83d-50a0104d06d9","Type":"ContainerStarted","Data":"173d03c6a95edb199581f0d5117497414c7ba5718e997ef08b4ebdf4fb61cb7a"} Nov 23 07:09:16 crc kubenswrapper[4906]: I1123 07:09:16.922843 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-slkpb" event={"ID":"0b28db04-7998-4e4c-80c2-dc2e246f55c5","Type":"ContainerStarted","Data":"531081772b215724b14739dbe8db3d3198c4edaa7161148e8553a2bddb146101"} Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.933480 4906 generic.go:334] "Generic (PLEG): container finished" podID="7e9fa61a-edd3-476d-8084-10653351773b" containerID="018decda61999cccfd06ac4b20b5448abba52ae9229f9dd58c47095fdc6d2e57" exitCode=0 Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.933590 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vzfd7" event={"ID":"7e9fa61a-edd3-476d-8084-10653351773b","Type":"ContainerDied","Data":"018decda61999cccfd06ac4b20b5448abba52ae9229f9dd58c47095fdc6d2e57"} Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.938905 4906 generic.go:334] "Generic (PLEG): container finished" podID="02f3febd-9b3c-4329-a767-16b438b44c91" containerID="64aa8715dc63cefdc0464b31e6dedda624baac63db9568706ac79df76a4f97a5" exitCode=0 Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.939030 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f60-account-create-ljk4z" event={"ID":"02f3febd-9b3c-4329-a767-16b438b44c91","Type":"ContainerDied","Data":"64aa8715dc63cefdc0464b31e6dedda624baac63db9568706ac79df76a4f97a5"} Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.940770 4906 generic.go:334] "Generic (PLEG): container finished" podID="2c2f5a87-926b-4d48-8e44-ec8dc87f8462" containerID="1964d4fe020c161bced8f66ff4439496702b1f867d376b389cd394821610ab26" exitCode=0 Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.940839 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a450-account-create-8nn5h" event={"ID":"2c2f5a87-926b-4d48-8e44-ec8dc87f8462","Type":"ContainerDied","Data":"1964d4fe020c161bced8f66ff4439496702b1f867d376b389cd394821610ab26"} Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.942780 4906 generic.go:334] "Generic (PLEG): container finished" podID="0b28db04-7998-4e4c-80c2-dc2e246f55c5" containerID="531081772b215724b14739dbe8db3d3198c4edaa7161148e8553a2bddb146101" exitCode=0 Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.942846 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-slkpb" event={"ID":"0b28db04-7998-4e4c-80c2-dc2e246f55c5","Type":"ContainerDied","Data":"531081772b215724b14739dbe8db3d3198c4edaa7161148e8553a2bddb146101"} Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.944358 4906 generic.go:334] "Generic (PLEG): container finished" podID="aac4be73-123c-4818-b96d-0a6ea8898fcb" containerID="d6b6def31b94f53e43ab11d1f9b18fccdd68fc5016706a678167f411d20a9456" exitCode=0 Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.944409 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4vqvb" event={"ID":"aac4be73-123c-4818-b96d-0a6ea8898fcb","Type":"ContainerDied","Data":"d6b6def31b94f53e43ab11d1f9b18fccdd68fc5016706a678167f411d20a9456"} Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.954186 4906 generic.go:334] "Generic (PLEG): container finished" podID="0c44516c-f839-4d45-b83d-50a0104d06d9" containerID="173d03c6a95edb199581f0d5117497414c7ba5718e997ef08b4ebdf4fb61cb7a" exitCode=0 Nov 23 07:09:17 crc kubenswrapper[4906]: I1123 07:09:17.954247 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-46ad-account-create-wc7t2" event={"ID":"0c44516c-f839-4d45-b83d-50a0104d06d9","Type":"ContainerDied","Data":"173d03c6a95edb199581f0d5117497414c7ba5718e997ef08b4ebdf4fb61cb7a"} Nov 23 07:09:20 crc kubenswrapper[4906]: I1123 07:09:20.497933 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:20 crc kubenswrapper[4906]: I1123 07:09:20.562454 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-b6tpv"] Nov 23 07:09:20 crc kubenswrapper[4906]: I1123 07:09:20.562727 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" podUID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerName="dnsmasq-dns" containerID="cri-o://c22abeb905abc2533734bfb50f54b7f8670728d975ded0e99476ac3685b3c6dc" gracePeriod=10 Nov 23 07:09:20 crc kubenswrapper[4906]: I1123 07:09:20.981327 4906 generic.go:334] "Generic (PLEG): container finished" podID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerID="c22abeb905abc2533734bfb50f54b7f8670728d975ded0e99476ac3685b3c6dc" exitCode=0 Nov 23 07:09:20 crc kubenswrapper[4906]: I1123 07:09:20.981376 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" event={"ID":"0cd9d13e-9b11-44a7-9605-5e6d85825993","Type":"ContainerDied","Data":"c22abeb905abc2533734bfb50f54b7f8670728d975ded0e99476ac3685b3c6dc"} Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.540554 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.556622 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.594735 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.602473 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.618039 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.635597 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.661918 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9fa61a-edd3-476d-8084-10653351773b-operator-scripts\") pod \"7e9fa61a-edd3-476d-8084-10653351773b\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.662025 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5w6p\" (UniqueName: \"kubernetes.io/projected/7e9fa61a-edd3-476d-8084-10653351773b-kube-api-access-l5w6p\") pod \"7e9fa61a-edd3-476d-8084-10653351773b\" (UID: \"7e9fa61a-edd3-476d-8084-10653351773b\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.662188 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-operator-scripts\") pod \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.662217 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f2xl\" (UniqueName: \"kubernetes.io/projected/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-kube-api-access-8f2xl\") pod \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\" (UID: \"2c2f5a87-926b-4d48-8e44-ec8dc87f8462\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.663523 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9fa61a-edd3-476d-8084-10653351773b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e9fa61a-edd3-476d-8084-10653351773b" (UID: "7e9fa61a-edd3-476d-8084-10653351773b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.664152 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2c2f5a87-926b-4d48-8e44-ec8dc87f8462" (UID: "2c2f5a87-926b-4d48-8e44-ec8dc87f8462"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.673116 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9fa61a-edd3-476d-8084-10653351773b-kube-api-access-l5w6p" (OuterVolumeSpecName: "kube-api-access-l5w6p") pod "7e9fa61a-edd3-476d-8084-10653351773b" (UID: "7e9fa61a-edd3-476d-8084-10653351773b"). InnerVolumeSpecName "kube-api-access-l5w6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.674458 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-kube-api-access-8f2xl" (OuterVolumeSpecName: "kube-api-access-8f2xl") pod "2c2f5a87-926b-4d48-8e44-ec8dc87f8462" (UID: "2c2f5a87-926b-4d48-8e44-ec8dc87f8462"). InnerVolumeSpecName "kube-api-access-8f2xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.726642 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.763898 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c44516c-f839-4d45-b83d-50a0104d06d9-operator-scripts\") pod \"0c44516c-f839-4d45-b83d-50a0104d06d9\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.763975 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tkbs\" (UniqueName: \"kubernetes.io/projected/0c44516c-f839-4d45-b83d-50a0104d06d9-kube-api-access-4tkbs\") pod \"0c44516c-f839-4d45-b83d-50a0104d06d9\" (UID: \"0c44516c-f839-4d45-b83d-50a0104d06d9\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764035 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02f3febd-9b3c-4329-a767-16b438b44c91-operator-scripts\") pod \"02f3febd-9b3c-4329-a767-16b438b44c91\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764087 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtskw\" (UniqueName: \"kubernetes.io/projected/0b28db04-7998-4e4c-80c2-dc2e246f55c5-kube-api-access-xtskw\") pod \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764216 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b28db04-7998-4e4c-80c2-dc2e246f55c5-operator-scripts\") pod \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\" (UID: \"0b28db04-7998-4e4c-80c2-dc2e246f55c5\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764353 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac4be73-123c-4818-b96d-0a6ea8898fcb-operator-scripts\") pod \"aac4be73-123c-4818-b96d-0a6ea8898fcb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764455 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nr69\" (UniqueName: \"kubernetes.io/projected/aac4be73-123c-4818-b96d-0a6ea8898fcb-kube-api-access-2nr69\") pod \"aac4be73-123c-4818-b96d-0a6ea8898fcb\" (UID: \"aac4be73-123c-4818-b96d-0a6ea8898fcb\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764557 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gggdj\" (UniqueName: \"kubernetes.io/projected/02f3febd-9b3c-4329-a767-16b438b44c91-kube-api-access-gggdj\") pod \"02f3febd-9b3c-4329-a767-16b438b44c91\" (UID: \"02f3febd-9b3c-4329-a767-16b438b44c91\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764637 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02f3febd-9b3c-4329-a767-16b438b44c91-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "02f3febd-9b3c-4329-a767-16b438b44c91" (UID: "02f3febd-9b3c-4329-a767-16b438b44c91"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764487 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c44516c-f839-4d45-b83d-50a0104d06d9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c44516c-f839-4d45-b83d-50a0104d06d9" (UID: "0c44516c-f839-4d45-b83d-50a0104d06d9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.764771 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b28db04-7998-4e4c-80c2-dc2e246f55c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0b28db04-7998-4e4c-80c2-dc2e246f55c5" (UID: "0b28db04-7998-4e4c-80c2-dc2e246f55c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.765149 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aac4be73-123c-4818-b96d-0a6ea8898fcb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aac4be73-123c-4818-b96d-0a6ea8898fcb" (UID: "aac4be73-123c-4818-b96d-0a6ea8898fcb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.765894 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5w6p\" (UniqueName: \"kubernetes.io/projected/7e9fa61a-edd3-476d-8084-10653351773b-kube-api-access-l5w6p\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.765933 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c44516c-f839-4d45-b83d-50a0104d06d9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.765951 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02f3febd-9b3c-4329-a767-16b438b44c91-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.765964 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b28db04-7998-4e4c-80c2-dc2e246f55c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.765977 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aac4be73-123c-4818-b96d-0a6ea8898fcb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.765989 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.766001 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f2xl\" (UniqueName: \"kubernetes.io/projected/2c2f5a87-926b-4d48-8e44-ec8dc87f8462-kube-api-access-8f2xl\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.766015 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e9fa61a-edd3-476d-8084-10653351773b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.783346 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c44516c-f839-4d45-b83d-50a0104d06d9-kube-api-access-4tkbs" (OuterVolumeSpecName: "kube-api-access-4tkbs") pod "0c44516c-f839-4d45-b83d-50a0104d06d9" (UID: "0c44516c-f839-4d45-b83d-50a0104d06d9"). InnerVolumeSpecName "kube-api-access-4tkbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.783535 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b28db04-7998-4e4c-80c2-dc2e246f55c5-kube-api-access-xtskw" (OuterVolumeSpecName: "kube-api-access-xtskw") pod "0b28db04-7998-4e4c-80c2-dc2e246f55c5" (UID: "0b28db04-7998-4e4c-80c2-dc2e246f55c5"). InnerVolumeSpecName "kube-api-access-xtskw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.783531 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac4be73-123c-4818-b96d-0a6ea8898fcb-kube-api-access-2nr69" (OuterVolumeSpecName: "kube-api-access-2nr69") pod "aac4be73-123c-4818-b96d-0a6ea8898fcb" (UID: "aac4be73-123c-4818-b96d-0a6ea8898fcb"). InnerVolumeSpecName "kube-api-access-2nr69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.785508 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f3febd-9b3c-4329-a767-16b438b44c91-kube-api-access-gggdj" (OuterVolumeSpecName: "kube-api-access-gggdj") pod "02f3febd-9b3c-4329-a767-16b438b44c91" (UID: "02f3febd-9b3c-4329-a767-16b438b44c91"). InnerVolumeSpecName "kube-api-access-gggdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.867300 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-sb\") pod \"0cd9d13e-9b11-44a7-9605-5e6d85825993\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.867900 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-dns-svc\") pod \"0cd9d13e-9b11-44a7-9605-5e6d85825993\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.867975 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-config\") pod \"0cd9d13e-9b11-44a7-9605-5e6d85825993\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.868054 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r44b7\" (UniqueName: \"kubernetes.io/projected/0cd9d13e-9b11-44a7-9605-5e6d85825993-kube-api-access-r44b7\") pod \"0cd9d13e-9b11-44a7-9605-5e6d85825993\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.868122 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-nb\") pod \"0cd9d13e-9b11-44a7-9605-5e6d85825993\" (UID: \"0cd9d13e-9b11-44a7-9605-5e6d85825993\") " Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.868501 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tkbs\" (UniqueName: \"kubernetes.io/projected/0c44516c-f839-4d45-b83d-50a0104d06d9-kube-api-access-4tkbs\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.868520 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtskw\" (UniqueName: \"kubernetes.io/projected/0b28db04-7998-4e4c-80c2-dc2e246f55c5-kube-api-access-xtskw\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.868531 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nr69\" (UniqueName: \"kubernetes.io/projected/aac4be73-123c-4818-b96d-0a6ea8898fcb-kube-api-access-2nr69\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.868544 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gggdj\" (UniqueName: \"kubernetes.io/projected/02f3febd-9b3c-4329-a767-16b438b44c91-kube-api-access-gggdj\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.872109 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cd9d13e-9b11-44a7-9605-5e6d85825993-kube-api-access-r44b7" (OuterVolumeSpecName: "kube-api-access-r44b7") pod "0cd9d13e-9b11-44a7-9605-5e6d85825993" (UID: "0cd9d13e-9b11-44a7-9605-5e6d85825993"). InnerVolumeSpecName "kube-api-access-r44b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.913792 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-config" (OuterVolumeSpecName: "config") pod "0cd9d13e-9b11-44a7-9605-5e6d85825993" (UID: "0cd9d13e-9b11-44a7-9605-5e6d85825993"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.920731 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0cd9d13e-9b11-44a7-9605-5e6d85825993" (UID: "0cd9d13e-9b11-44a7-9605-5e6d85825993"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.930675 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0cd9d13e-9b11-44a7-9605-5e6d85825993" (UID: "0cd9d13e-9b11-44a7-9605-5e6d85825993"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.932148 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0cd9d13e-9b11-44a7-9605-5e6d85825993" (UID: "0cd9d13e-9b11-44a7-9605-5e6d85825993"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.971519 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.971589 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.971612 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r44b7\" (UniqueName: \"kubernetes.io/projected/0cd9d13e-9b11-44a7-9605-5e6d85825993-kube-api-access-r44b7\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.971645 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:22 crc kubenswrapper[4906]: I1123 07:09:22.971668 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cd9d13e-9b11-44a7-9605-5e6d85825993-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.011041 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a450-account-create-8nn5h" event={"ID":"2c2f5a87-926b-4d48-8e44-ec8dc87f8462","Type":"ContainerDied","Data":"3a7ba6fa7eaa4ccfc703b4704462f502740b8daaf5787cc5c410ff63a54492b7"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.011093 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a7ba6fa7eaa4ccfc703b4704462f502740b8daaf5787cc5c410ff63a54492b7" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.011183 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a450-account-create-8nn5h" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.019594 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-slkpb" event={"ID":"0b28db04-7998-4e4c-80c2-dc2e246f55c5","Type":"ContainerDied","Data":"4197fcea1f9b42c20cf7caa24964a54ff13d7bb534ed6f67d067d0a9f3547fe8"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.019615 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-slkpb" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.019897 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4197fcea1f9b42c20cf7caa24964a54ff13d7bb534ed6f67d067d0a9f3547fe8" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.021490 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4vqvb" event={"ID":"aac4be73-123c-4818-b96d-0a6ea8898fcb","Type":"ContainerDied","Data":"9f2b4a5ca47c646e337d88bfac4adea25e7922173e247b5606a16f2311ffc372"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.021649 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f2b4a5ca47c646e337d88bfac4adea25e7922173e247b5606a16f2311ffc372" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.021508 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4vqvb" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.024453 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c7r8h" event={"ID":"bb7e51ef-70f7-4fb2-a1ef-31a109125528","Type":"ContainerStarted","Data":"feeab8377b8ce338f687bcf8f59f53e807ba5d4b1a8a47037bf6e81ca91565dc"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.034039 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" event={"ID":"0cd9d13e-9b11-44a7-9605-5e6d85825993","Type":"ContainerDied","Data":"1c380d7549e14ce1101d48d59433d0e8dc07b0e687a4598393eb74c8475a2b7e"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.034127 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-b6tpv" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.034172 4906 scope.go:117] "RemoveContainer" containerID="c22abeb905abc2533734bfb50f54b7f8670728d975ded0e99476ac3685b3c6dc" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.036549 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-46ad-account-create-wc7t2" event={"ID":"0c44516c-f839-4d45-b83d-50a0104d06d9","Type":"ContainerDied","Data":"297e93b0f85689e56d1ced37a96a144993d3cd59153fafda192b6b95bed345cf"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.036613 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="297e93b0f85689e56d1ced37a96a144993d3cd59153fafda192b6b95bed345cf" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.036619 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-46ad-account-create-wc7t2" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.038752 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vzfd7" event={"ID":"7e9fa61a-edd3-476d-8084-10653351773b","Type":"ContainerDied","Data":"0f7357163617a64ee425d2b5481721b4b3e0a9c5f0739da3add0583400aaf2a0"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.038781 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f7357163617a64ee425d2b5481721b4b3e0a9c5f0739da3add0583400aaf2a0" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.038827 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzfd7" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.045480 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9f60-account-create-ljk4z" event={"ID":"02f3febd-9b3c-4329-a767-16b438b44c91","Type":"ContainerDied","Data":"8e120d61e2177096d7bd7af35a9d0b9a7471d4cc69e23f1361986864ccca96c1"} Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.045543 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e120d61e2177096d7bd7af35a9d0b9a7471d4cc69e23f1361986864ccca96c1" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.045567 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9f60-account-create-ljk4z" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.057088 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-c7r8h" podStartSLOduration=2.90214234 podStartE2EDuration="11.057064933s" podCreationTimestamp="2025-11-23 07:09:12 +0000 UTC" firstStartedPulling="2025-11-23 07:09:14.244355565 +0000 UTC m=+1169.757746868" lastFinishedPulling="2025-11-23 07:09:22.399278158 +0000 UTC m=+1177.912669461" observedRunningTime="2025-11-23 07:09:23.051952799 +0000 UTC m=+1178.565344102" watchObservedRunningTime="2025-11-23 07:09:23.057064933 +0000 UTC m=+1178.570456236" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.086749 4906 scope.go:117] "RemoveContainer" containerID="10f7a5e65693fad4378780c52506ab5c80d765de6888636c2b598b1ec596f52c" Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.136110 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-b6tpv"] Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.143973 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-b6tpv"] Nov 23 07:09:23 crc kubenswrapper[4906]: I1123 07:09:23.372792 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cd9d13e-9b11-44a7-9605-5e6d85825993" path="/var/lib/kubelet/pods/0cd9d13e-9b11-44a7-9605-5e6d85825993/volumes" Nov 23 07:09:26 crc kubenswrapper[4906]: I1123 07:09:26.083280 4906 generic.go:334] "Generic (PLEG): container finished" podID="bb7e51ef-70f7-4fb2-a1ef-31a109125528" containerID="feeab8377b8ce338f687bcf8f59f53e807ba5d4b1a8a47037bf6e81ca91565dc" exitCode=0 Nov 23 07:09:26 crc kubenswrapper[4906]: I1123 07:09:26.083343 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c7r8h" event={"ID":"bb7e51ef-70f7-4fb2-a1ef-31a109125528","Type":"ContainerDied","Data":"feeab8377b8ce338f687bcf8f59f53e807ba5d4b1a8a47037bf6e81ca91565dc"} Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.494144 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.668767 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-combined-ca-bundle\") pod \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.668853 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-config-data\") pod \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.668967 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vn8c\" (UniqueName: \"kubernetes.io/projected/bb7e51ef-70f7-4fb2-a1ef-31a109125528-kube-api-access-9vn8c\") pod \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\" (UID: \"bb7e51ef-70f7-4fb2-a1ef-31a109125528\") " Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.678962 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7e51ef-70f7-4fb2-a1ef-31a109125528-kube-api-access-9vn8c" (OuterVolumeSpecName: "kube-api-access-9vn8c") pod "bb7e51ef-70f7-4fb2-a1ef-31a109125528" (UID: "bb7e51ef-70f7-4fb2-a1ef-31a109125528"). InnerVolumeSpecName "kube-api-access-9vn8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.717193 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb7e51ef-70f7-4fb2-a1ef-31a109125528" (UID: "bb7e51ef-70f7-4fb2-a1ef-31a109125528"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.742500 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-config-data" (OuterVolumeSpecName: "config-data") pod "bb7e51ef-70f7-4fb2-a1ef-31a109125528" (UID: "bb7e51ef-70f7-4fb2-a1ef-31a109125528"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.771349 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.771404 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7e51ef-70f7-4fb2-a1ef-31a109125528-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:27 crc kubenswrapper[4906]: I1123 07:09:27.771423 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vn8c\" (UniqueName: \"kubernetes.io/projected/bb7e51ef-70f7-4fb2-a1ef-31a109125528-kube-api-access-9vn8c\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.103159 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c7r8h" event={"ID":"bb7e51ef-70f7-4fb2-a1ef-31a109125528","Type":"ContainerDied","Data":"53e291ba23eae571244763f110de9045f9f507418c7aadbe815d57663611b1eb"} Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.103230 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53e291ba23eae571244763f110de9045f9f507418c7aadbe815d57663611b1eb" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.103596 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c7r8h" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.393930 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-d7bxl"] Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394272 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerName="init" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394290 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerName="init" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394307 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9fa61a-edd3-476d-8084-10653351773b" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394313 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9fa61a-edd3-476d-8084-10653351773b" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394326 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerName="dnsmasq-dns" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394334 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerName="dnsmasq-dns" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394343 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac4be73-123c-4818-b96d-0a6ea8898fcb" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394350 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac4be73-123c-4818-b96d-0a6ea8898fcb" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394360 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b28db04-7998-4e4c-80c2-dc2e246f55c5" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394365 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b28db04-7998-4e4c-80c2-dc2e246f55c5" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394383 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2f5a87-926b-4d48-8e44-ec8dc87f8462" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394390 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2f5a87-926b-4d48-8e44-ec8dc87f8462" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394405 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c44516c-f839-4d45-b83d-50a0104d06d9" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394411 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c44516c-f839-4d45-b83d-50a0104d06d9" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394423 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f3febd-9b3c-4329-a767-16b438b44c91" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394431 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f3febd-9b3c-4329-a767-16b438b44c91" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: E1123 07:09:28.394448 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7e51ef-70f7-4fb2-a1ef-31a109125528" containerName="keystone-db-sync" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394453 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7e51ef-70f7-4fb2-a1ef-31a109125528" containerName="keystone-db-sync" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394595 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f3febd-9b3c-4329-a767-16b438b44c91" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394611 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7e51ef-70f7-4fb2-a1ef-31a109125528" containerName="keystone-db-sync" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394621 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cd9d13e-9b11-44a7-9605-5e6d85825993" containerName="dnsmasq-dns" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394633 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9fa61a-edd3-476d-8084-10653351773b" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394643 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c44516c-f839-4d45-b83d-50a0104d06d9" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394650 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b28db04-7998-4e4c-80c2-dc2e246f55c5" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394661 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c2f5a87-926b-4d48-8e44-ec8dc87f8462" containerName="mariadb-account-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.394670 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac4be73-123c-4818-b96d-0a6ea8898fcb" containerName="mariadb-database-create" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.396741 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.413329 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-d7bxl"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.427443 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9qdmg"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.429067 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.431248 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.431255 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.431335 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.433044 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.433217 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nk8ll" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.455526 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9qdmg"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.489434 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-svc\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.489502 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.489543 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.489585 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j78b5\" (UniqueName: \"kubernetes.io/projected/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-kube-api-access-j78b5\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.489628 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-swift-storage-0\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.489748 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-config\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.589620 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-ltbgj"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.591232 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593323 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-config\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593351 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8sd2\" (UniqueName: \"kubernetes.io/projected/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-kube-api-access-j8sd2\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593396 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-svc\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593413 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-fernet-keys\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593437 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593476 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593506 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-credential-keys\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593521 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-combined-ca-bundle\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593535 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-scripts\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593558 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j78b5\" (UniqueName: \"kubernetes.io/projected/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-kube-api-access-j78b5\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593581 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-swift-storage-0\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.593614 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-config-data\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.594728 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-config\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.595315 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-svc\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.596153 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.596165 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-swift-storage-0\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.596725 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.597262 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.597507 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.597690 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4555g" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.606523 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-6p4bd"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.609715 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.612901 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.613241 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.618521 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ltbgj"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.620757 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-h6n2h" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.650059 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6p4bd"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.686462 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j78b5\" (UniqueName: \"kubernetes.io/projected/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-kube-api-access-j78b5\") pod \"dnsmasq-dns-7dbf8bff67-d7bxl\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.695534 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-config-data\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.695606 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-fernet-keys\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.695644 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-config\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.695671 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2d3d804-9bab-439f-94c6-23e1546aaad7-etc-machine-id\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.695888 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-combined-ca-bundle\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.695936 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-db-sync-config-data\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.695974 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrsg9\" (UniqueName: \"kubernetes.io/projected/0cc86f52-0139-430f-9803-06b714d9fc7e-kube-api-access-lrsg9\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696008 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-credential-keys\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696028 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-combined-ca-bundle\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696050 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-scripts\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696097 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjxdv\" (UniqueName: \"kubernetes.io/projected/d2d3d804-9bab-439f-94c6-23e1546aaad7-kube-api-access-zjxdv\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696128 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-scripts\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696147 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-combined-ca-bundle\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696249 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-config-data\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.696304 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8sd2\" (UniqueName: \"kubernetes.io/projected/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-kube-api-access-j8sd2\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.700107 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-combined-ca-bundle\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.700353 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-credential-keys\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.701214 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-fernet-keys\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.704634 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-config-data\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.709092 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-scripts\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.722215 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.784608 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8sd2\" (UniqueName: \"kubernetes.io/projected/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-kube-api-access-j8sd2\") pod \"keystone-bootstrap-9qdmg\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.785842 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.793210 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.795243 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798566 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-combined-ca-bundle\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798609 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-db-sync-config-data\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798632 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrsg9\" (UniqueName: \"kubernetes.io/projected/0cc86f52-0139-430f-9803-06b714d9fc7e-kube-api-access-lrsg9\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798671 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjxdv\" (UniqueName: \"kubernetes.io/projected/d2d3d804-9bab-439f-94c6-23e1546aaad7-kube-api-access-zjxdv\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798708 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-combined-ca-bundle\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798726 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-scripts\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798796 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-config-data\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798822 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-config\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798841 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2d3d804-9bab-439f-94c6-23e1546aaad7-etc-machine-id\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.798948 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2d3d804-9bab-439f-94c6-23e1546aaad7-etc-machine-id\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.801899 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.802606 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.805168 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-combined-ca-bundle\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.806697 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-combined-ca-bundle\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.814791 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-config-data\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.823703 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-db-sync-config-data\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.831732 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-scripts\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.845151 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-config\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.851138 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrsg9\" (UniqueName: \"kubernetes.io/projected/0cc86f52-0139-430f-9803-06b714d9fc7e-kube-api-access-lrsg9\") pod \"neutron-db-sync-ltbgj\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.878056 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-d7bxl"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.899879 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.899921 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-log-httpd\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.899952 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-run-httpd\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.899989 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8w7w\" (UniqueName: \"kubernetes.io/projected/865cb0c4-2164-4c89-9153-20f19e15db0c-kube-api-access-z8w7w\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.900010 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-scripts\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.900033 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.900081 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-config-data\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.931714 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.959339 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjxdv\" (UniqueName: \"kubernetes.io/projected/d2d3d804-9bab-439f-94c6-23e1546aaad7-kube-api-access-zjxdv\") pod \"cinder-db-sync-6p4bd\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.969199 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6wvm5"] Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.970773 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.976830 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.977081 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 23 07:09:28 crc kubenswrapper[4906]: I1123 07:09:28.991902 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-59ssp" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.002546 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.002582 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-log-httpd\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.002617 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-run-httpd\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.002647 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8w7w\" (UniqueName: \"kubernetes.io/projected/865cb0c4-2164-4c89-9153-20f19e15db0c-kube-api-access-z8w7w\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.002671 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-scripts\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.002717 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.002761 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-config-data\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.005205 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-run-httpd\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.006141 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-log-httpd\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.015214 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.015260 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6wvm5"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.015232 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-config-data\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.016673 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-scripts\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.033824 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.044386 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8w7w\" (UniqueName: \"kubernetes.io/projected/865cb0c4-2164-4c89-9153-20f19e15db0c-kube-api-access-z8w7w\") pod \"ceilometer-0\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.048986 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9l4vd"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.050343 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.056370 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b2ntq" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.056617 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.059229 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.104666 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tgkp\" (UniqueName: \"kubernetes.io/projected/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-kube-api-access-7tgkp\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.104755 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-logs\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.104780 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-scripts\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.104862 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-config-data\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.104897 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-combined-ca-bundle\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.122157 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9l4vd"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.135912 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-595vk"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.137494 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.140040 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-595vk"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206075 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-logs\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206124 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-scripts\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206162 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-db-sync-config-data\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206183 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206203 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-config\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206226 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf5w5\" (UniqueName: \"kubernetes.io/projected/1e49ca90-ea62-4f3b-88de-bc876ef28d65-kube-api-access-xf5w5\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206256 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2smc\" (UniqueName: \"kubernetes.io/projected/5584696b-3782-4db6-997e-9f691f70d05a-kube-api-access-d2smc\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206297 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206323 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-config-data\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206353 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206372 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-combined-ca-bundle\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206408 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206433 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-combined-ca-bundle\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.206459 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tgkp\" (UniqueName: \"kubernetes.io/projected/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-kube-api-access-7tgkp\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.213172 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-logs\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.229772 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-config-data\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.264547 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-combined-ca-bundle\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.266694 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-scripts\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.268175 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.269927 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.284052 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tgkp\" (UniqueName: \"kubernetes.io/projected/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-kube-api-access-7tgkp\") pod \"placement-db-sync-6wvm5\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.308488 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-db-sync-config-data\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.308531 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.308552 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-config\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.308572 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf5w5\" (UniqueName: \"kubernetes.io/projected/1e49ca90-ea62-4f3b-88de-bc876ef28d65-kube-api-access-xf5w5\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.308603 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2smc\" (UniqueName: \"kubernetes.io/projected/5584696b-3782-4db6-997e-9f691f70d05a-kube-api-access-d2smc\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.308641 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.309547 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-config\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.309965 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.310024 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.310046 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-combined-ca-bundle\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.310934 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.311633 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.312264 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.312703 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.313574 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-db-sync-config-data\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.314049 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-combined-ca-bundle\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.329032 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6wvm5" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.330015 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf5w5\" (UniqueName: \"kubernetes.io/projected/1e49ca90-ea62-4f3b-88de-bc876ef28d65-kube-api-access-xf5w5\") pod \"dnsmasq-dns-76c58b6d97-595vk\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.332739 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2smc\" (UniqueName: \"kubernetes.io/projected/5584696b-3782-4db6-997e-9f691f70d05a-kube-api-access-d2smc\") pod \"barbican-db-sync-9l4vd\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.457113 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.493669 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.542594 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.544626 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.551447 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5v68l" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.556624 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.556748 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.557013 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.566158 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.619667 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-d7bxl"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624163 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624227 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624263 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624289 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624324 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-logs\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624366 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh65l\" (UniqueName: \"kubernetes.io/projected/d5770119-9b13-474e-b564-a183d2dd9bb5-kube-api-access-bh65l\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624394 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.624437 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.696531 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.701856 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.705324 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.705685 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.717134 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725774 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-logs\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725816 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725868 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh65l\" (UniqueName: \"kubernetes.io/projected/d5770119-9b13-474e-b564-a183d2dd9bb5-kube-api-access-bh65l\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725889 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725919 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725958 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725974 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.725999 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726018 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726049 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726067 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726088 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726119 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726142 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726161 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-logs\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726179 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvqp2\" (UniqueName: \"kubernetes.io/projected/c5878bc8-5cd8-4a31-8c7e-530618fb4848-kube-api-access-wvqp2\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726336 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-logs\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.726905 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.729747 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.734632 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.737974 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.760190 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh65l\" (UniqueName: \"kubernetes.io/projected/d5770119-9b13-474e-b564-a183d2dd9bb5-kube-api-access-bh65l\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.763168 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.779418 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.783909 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.813551 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ltbgj"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.837387 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.838298 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.838379 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.838581 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.838636 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.838812 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-logs\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.838847 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvqp2\" (UniqueName: \"kubernetes.io/projected/c5878bc8-5cd8-4a31-8c7e-530618fb4848-kube-api-access-wvqp2\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.838986 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.840119 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.849116 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-logs\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.849469 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.849649 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.859670 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9qdmg"] Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.860973 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.861733 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.865281 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvqp2\" (UniqueName: \"kubernetes.io/projected/c5878bc8-5cd8-4a31-8c7e-530618fb4848-kube-api-access-wvqp2\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.866305 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.880589 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.894265 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.941004 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:29 crc kubenswrapper[4906]: I1123 07:09:29.979038 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6p4bd"] Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.194162 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.210375 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6wvm5"] Nov 23 07:09:30 crc kubenswrapper[4906]: W1123 07:09:30.236699 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod865cb0c4_2164_4c89_9153_20f19e15db0c.slice/crio-4503eede6f143cfc69698a84c0e9db60a5906fa8a4e8c8427e2d1a838c44421e WatchSource:0}: Error finding container 4503eede6f143cfc69698a84c0e9db60a5906fa8a4e8c8427e2d1a838c44421e: Status 404 returned error can't find the container with id 4503eede6f143cfc69698a84c0e9db60a5906fa8a4e8c8427e2d1a838c44421e Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.309171 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerStarted","Data":"4503eede6f143cfc69698a84c0e9db60a5906fa8a4e8c8427e2d1a838c44421e"} Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.311606 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" event={"ID":"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05","Type":"ContainerStarted","Data":"a2ca0ea2d00d5d91256f951cd2496d43e1ea850dff002ebb074d87e2efde27db"} Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.312519 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6p4bd" event={"ID":"d2d3d804-9bab-439f-94c6-23e1546aaad7","Type":"ContainerStarted","Data":"4037c6ff51a31b197f817c2f158c2b2dc19ab3ae2763acb61c021a40cf648165"} Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.314618 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qdmg" event={"ID":"de58c3f3-2425-4137-8c3c-9871d1fbd9c1","Type":"ContainerStarted","Data":"55b22750cb735ba0af2f22a508eacdaaed269239b61910400c6ad3c92e46ea23"} Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.323420 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6wvm5" event={"ID":"8f23c4a7-b2b9-4ef9-876b-5dd95f486333","Type":"ContainerStarted","Data":"e65b6796fc5a43c9b616dbcb5c30b54532e6c9c0e24b12e56d3ccbd0811e8029"} Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.324797 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9l4vd"] Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.329665 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ltbgj" event={"ID":"0cc86f52-0139-430f-9803-06b714d9fc7e","Type":"ContainerStarted","Data":"ca0dfaa3b3b681f02dcd40ec304d45c72edc9225c7d05e2761efe5001a5e6a60"} Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.333179 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-595vk"] Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.441716 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:30 crc kubenswrapper[4906]: I1123 07:09:30.750407 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:30 crc kubenswrapper[4906]: W1123 07:09:30.757843 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5878bc8_5cd8_4a31_8c7e_530618fb4848.slice/crio-b306a53a03c57a831d103e1e5c79daf89992de0c1c500b7fc92d3dcab9965884 WatchSource:0}: Error finding container b306a53a03c57a831d103e1e5c79daf89992de0c1c500b7fc92d3dcab9965884: Status 404 returned error can't find the container with id b306a53a03c57a831d103e1e5c79daf89992de0c1c500b7fc92d3dcab9965884 Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.373454 4906 generic.go:334] "Generic (PLEG): container finished" podID="e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" containerID="d498b06b5e23d63965b93984031fc20bb2c6732f60513c6add93ee9d96675797" exitCode=0 Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.385108 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5770119-9b13-474e-b564-a183d2dd9bb5","Type":"ContainerStarted","Data":"fa1388c93c027b9be112c68c4a139311a116f521b2e5437a02612d604c3d0784"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.385171 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ltbgj" event={"ID":"0cc86f52-0139-430f-9803-06b714d9fc7e","Type":"ContainerStarted","Data":"739fbb582901e5e7181fb6829f18d4295a29898fbc690f867445b5383dea4f42"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.385185 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5878bc8-5cd8-4a31-8c7e-530618fb4848","Type":"ContainerStarted","Data":"b306a53a03c57a831d103e1e5c79daf89992de0c1c500b7fc92d3dcab9965884"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.385275 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" event={"ID":"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05","Type":"ContainerDied","Data":"d498b06b5e23d63965b93984031fc20bb2c6732f60513c6add93ee9d96675797"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.385302 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" event={"ID":"1e49ca90-ea62-4f3b-88de-bc876ef28d65","Type":"ContainerStarted","Data":"b1307d8fa8d02b6b3c9334b2d3d293699ee2cb419fb6b42499ad3b0351985caa"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.385315 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" event={"ID":"1e49ca90-ea62-4f3b-88de-bc876ef28d65","Type":"ContainerStarted","Data":"fc438511ee7772182cc30c2f3d1b3f0e0e3b6a741fa3abf59d3cdc70ef8cdeb8"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.386362 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qdmg" event={"ID":"de58c3f3-2425-4137-8c3c-9871d1fbd9c1","Type":"ContainerStarted","Data":"cbd45e5087e3ff34aceaedab7642e3f533b4b897abad75afa1f69b5f5f11f37f"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.388494 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9l4vd" event={"ID":"5584696b-3782-4db6-997e-9f691f70d05a","Type":"ContainerStarted","Data":"6e20a48620a788b724cb505b39fd8e4ce789fc5289f259790e1fa5ead16977ff"} Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.398125 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-ltbgj" podStartSLOduration=3.3981017749999998 podStartE2EDuration="3.398101775s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:31.386312438 +0000 UTC m=+1186.899703781" watchObservedRunningTime="2025-11-23 07:09:31.398101775 +0000 UTC m=+1186.911493088" Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.439647 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9qdmg" podStartSLOduration=3.43962222 podStartE2EDuration="3.43962222s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:31.435783049 +0000 UTC m=+1186.949174352" watchObservedRunningTime="2025-11-23 07:09:31.43962222 +0000 UTC m=+1186.953013523" Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.644563 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.748143 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.800150 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:31 crc kubenswrapper[4906]: I1123 07:09:31.962899 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.022399 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-swift-storage-0\") pod \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.022815 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j78b5\" (UniqueName: \"kubernetes.io/projected/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-kube-api-access-j78b5\") pod \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.022961 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-svc\") pod \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.023042 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-nb\") pod \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.023222 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-config\") pod \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.023251 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-sb\") pod \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\" (UID: \"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05\") " Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.050163 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-kube-api-access-j78b5" (OuterVolumeSpecName: "kube-api-access-j78b5") pod "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" (UID: "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05"). InnerVolumeSpecName "kube-api-access-j78b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.054210 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" (UID: "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.059536 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" (UID: "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.070429 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" (UID: "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.087056 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" (UID: "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.096231 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-config" (OuterVolumeSpecName: "config") pod "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" (UID: "e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.146751 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.146794 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j78b5\" (UniqueName: \"kubernetes.io/projected/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-kube-api-access-j78b5\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.146808 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.146816 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.146826 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.146834 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.471117 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.471107 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbf8bff67-d7bxl" event={"ID":"e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05","Type":"ContainerDied","Data":"a2ca0ea2d00d5d91256f951cd2496d43e1ea850dff002ebb074d87e2efde27db"} Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.471634 4906 scope.go:117] "RemoveContainer" containerID="d498b06b5e23d63965b93984031fc20bb2c6732f60513c6add93ee9d96675797" Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.494403 4906 generic.go:334] "Generic (PLEG): container finished" podID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerID="b1307d8fa8d02b6b3c9334b2d3d293699ee2cb419fb6b42499ad3b0351985caa" exitCode=0 Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.494644 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" event={"ID":"1e49ca90-ea62-4f3b-88de-bc876ef28d65","Type":"ContainerDied","Data":"b1307d8fa8d02b6b3c9334b2d3d293699ee2cb419fb6b42499ad3b0351985caa"} Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.519966 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5770119-9b13-474e-b564-a183d2dd9bb5","Type":"ContainerStarted","Data":"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc"} Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.555850 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-d7bxl"] Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.556428 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5878bc8-5cd8-4a31-8c7e-530618fb4848","Type":"ContainerStarted","Data":"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e"} Nov 23 07:09:32 crc kubenswrapper[4906]: I1123 07:09:32.570496 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-d7bxl"] Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.368946 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" path="/var/lib/kubelet/pods/e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05/volumes" Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.570281 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" event={"ID":"1e49ca90-ea62-4f3b-88de-bc876ef28d65","Type":"ContainerStarted","Data":"bdc38c6e92d3fc410dab382cf993c78b39f4c8731349dc9479864fc24e34c033"} Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.570396 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.575013 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5770119-9b13-474e-b564-a183d2dd9bb5","Type":"ContainerStarted","Data":"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742"} Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.577440 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5878bc8-5cd8-4a31-8c7e-530618fb4848","Type":"ContainerStarted","Data":"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367"} Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.577602 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-log" containerID="cri-o://0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e" gracePeriod=30 Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.577697 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-httpd" containerID="cri-o://f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367" gracePeriod=30 Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.582716 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-httpd" containerID="cri-o://60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742" gracePeriod=30 Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.582728 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-log" containerID="cri-o://8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc" gracePeriod=30 Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.599324 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" podStartSLOduration=5.599298818 podStartE2EDuration="5.599298818s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:33.595854228 +0000 UTC m=+1189.109245521" watchObservedRunningTime="2025-11-23 07:09:33.599298818 +0000 UTC m=+1189.112690121" Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.640704 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.640663428 podStartE2EDuration="5.640663428s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:33.629582819 +0000 UTC m=+1189.142974122" watchObservedRunningTime="2025-11-23 07:09:33.640663428 +0000 UTC m=+1189.154054721" Nov 23 07:09:33 crc kubenswrapper[4906]: I1123 07:09:33.652459 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.652437295 podStartE2EDuration="5.652437295s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:33.649524859 +0000 UTC m=+1189.162916162" watchObservedRunningTime="2025-11-23 07:09:33.652437295 +0000 UTC m=+1189.165828598" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.347067 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.404979 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-logs\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.405040 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-scripts\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.405061 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-config-data\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.405195 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.405248 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-httpd-run\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.405278 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvqp2\" (UniqueName: \"kubernetes.io/projected/c5878bc8-5cd8-4a31-8c7e-530618fb4848-kube-api-access-wvqp2\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.405302 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-combined-ca-bundle\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.405380 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-internal-tls-certs\") pod \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\" (UID: \"c5878bc8-5cd8-4a31-8c7e-530618fb4848\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.409210 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.412785 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-logs" (OuterVolumeSpecName: "logs") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.416835 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5878bc8-5cd8-4a31-8c7e-530618fb4848-kube-api-access-wvqp2" (OuterVolumeSpecName: "kube-api-access-wvqp2") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "kube-api-access-wvqp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.416955 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.417097 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-scripts" (OuterVolumeSpecName: "scripts") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.439940 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.461855 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.465423 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.474555 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-config-data" (OuterVolumeSpecName: "config-data") pod "c5878bc8-5cd8-4a31-8c7e-530618fb4848" (UID: "c5878bc8-5cd8-4a31-8c7e-530618fb4848"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.508314 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-httpd-run\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.508403 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-combined-ca-bundle\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.508483 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-config-data\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.508839 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-logs\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.508891 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh65l\" (UniqueName: \"kubernetes.io/projected/d5770119-9b13-474e-b564-a183d2dd9bb5-kube-api-access-bh65l\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.508909 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509006 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-scripts\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509075 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-public-tls-certs\") pod \"d5770119-9b13-474e-b564-a183d2dd9bb5\" (UID: \"d5770119-9b13-474e-b564-a183d2dd9bb5\") " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509242 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509448 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509460 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvqp2\" (UniqueName: \"kubernetes.io/projected/c5878bc8-5cd8-4a31-8c7e-530618fb4848-kube-api-access-wvqp2\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509474 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509483 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509493 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5878bc8-5cd8-4a31-8c7e-530618fb4848-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509501 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509512 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5878bc8-5cd8-4a31-8c7e-530618fb4848-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509532 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509540 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.509566 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-logs" (OuterVolumeSpecName: "logs") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.518431 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.518722 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-scripts" (OuterVolumeSpecName: "scripts") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.524168 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5770119-9b13-474e-b564-a183d2dd9bb5-kube-api-access-bh65l" (OuterVolumeSpecName: "kube-api-access-bh65l") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "kube-api-access-bh65l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.542500 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.556538 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.568227 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-config-data" (OuterVolumeSpecName: "config-data") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.587266 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d5770119-9b13-474e-b564-a183d2dd9bb5" (UID: "d5770119-9b13-474e-b564-a183d2dd9bb5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.595826 4906 generic.go:334] "Generic (PLEG): container finished" podID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerID="60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742" exitCode=0 Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.595884 4906 generic.go:334] "Generic (PLEG): container finished" podID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerID="8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc" exitCode=143 Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.595938 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.595959 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5770119-9b13-474e-b564-a183d2dd9bb5","Type":"ContainerDied","Data":"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742"} Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.595993 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5770119-9b13-474e-b564-a183d2dd9bb5","Type":"ContainerDied","Data":"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc"} Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.596003 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5770119-9b13-474e-b564-a183d2dd9bb5","Type":"ContainerDied","Data":"fa1388c93c027b9be112c68c4a139311a116f521b2e5437a02612d604c3d0784"} Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.596038 4906 scope.go:117] "RemoveContainer" containerID="60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.602701 4906 generic.go:334] "Generic (PLEG): container finished" podID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerID="f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367" exitCode=0 Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.602724 4906 generic.go:334] "Generic (PLEG): container finished" podID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerID="0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e" exitCode=143 Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.602773 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.602815 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5878bc8-5cd8-4a31-8c7e-530618fb4848","Type":"ContainerDied","Data":"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367"} Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.602896 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5878bc8-5cd8-4a31-8c7e-530618fb4848","Type":"ContainerDied","Data":"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e"} Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.602912 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5878bc8-5cd8-4a31-8c7e-530618fb4848","Type":"ContainerDied","Data":"b306a53a03c57a831d103e1e5c79daf89992de0c1c500b7fc92d3dcab9965884"} Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611828 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611881 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5770119-9b13-474e-b564-a183d2dd9bb5-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611892 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh65l\" (UniqueName: \"kubernetes.io/projected/d5770119-9b13-474e-b564-a183d2dd9bb5-kube-api-access-bh65l\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611940 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611951 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611960 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611969 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.611980 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5770119-9b13-474e-b564-a183d2dd9bb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.639154 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.644208 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.652076 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.657874 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: E1123 07:09:34.658245 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" containerName="init" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.658265 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" containerName="init" Nov 23 07:09:34 crc kubenswrapper[4906]: E1123 07:09:34.658281 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-httpd" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.658287 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-httpd" Nov 23 07:09:34 crc kubenswrapper[4906]: E1123 07:09:34.658304 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-log" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659531 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-log" Nov 23 07:09:34 crc kubenswrapper[4906]: E1123 07:09:34.659554 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-log" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659562 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-log" Nov 23 07:09:34 crc kubenswrapper[4906]: E1123 07:09:34.659583 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-httpd" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659590 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-httpd" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659790 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3efd8f9-4f6f-4f6c-bb0c-06a97a86ad05" containerName="init" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659805 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-httpd" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659819 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-log" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659842 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" containerName="glance-httpd" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.659850 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" containerName="glance-log" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.661264 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.668446 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5v68l" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.668557 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.668657 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.669073 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.682250 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.716975 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-logs\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717027 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-config-data\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717069 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717141 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh4pc\" (UniqueName: \"kubernetes.io/projected/e6209532-4ce6-482f-9c69-021b4ec0b682-kube-api-access-wh4pc\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717167 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717237 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-scripts\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717280 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717317 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.717920 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.730118 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.737471 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.742975 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.745383 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.748357 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.752293 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.752460 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.772894 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.818755 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh4pc\" (UniqueName: \"kubernetes.io/projected/e6209532-4ce6-482f-9c69-021b4ec0b682-kube-api-access-wh4pc\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.818815 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.818863 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-scripts\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.818907 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.818961 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-config-data\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.818978 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-logs\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.819002 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.820083 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.820927 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-logs\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.823253 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.824793 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-config-data\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.826309 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-scripts\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.827050 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.837466 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh4pc\" (UniqueName: \"kubernetes.io/projected/e6209532-4ce6-482f-9c69-021b4ec0b682-kube-api-access-wh4pc\") pod \"glance-default-external-api-0\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " pod="openstack/glance-default-external-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921081 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921152 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921176 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921233 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-logs\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921265 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921302 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921345 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcmpg\" (UniqueName: \"kubernetes.io/projected/9e594b90-2851-4374-8913-08103bb5065e-kube-api-access-zcmpg\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.921364 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:34 crc kubenswrapper[4906]: I1123 07:09:34.993933 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.022608 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.022752 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.022782 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.022849 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-logs\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.022887 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.022930 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.022986 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcmpg\" (UniqueName: \"kubernetes.io/projected/9e594b90-2851-4374-8913-08103bb5065e-kube-api-access-zcmpg\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.023013 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.023768 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-logs\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.023814 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.024206 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.026835 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.028237 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.036020 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.039386 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.041816 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcmpg\" (UniqueName: \"kubernetes.io/projected/9e594b90-2851-4374-8913-08103bb5065e-kube-api-access-zcmpg\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.059010 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.198654 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.377583 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5878bc8-5cd8-4a31-8c7e-530618fb4848" path="/var/lib/kubelet/pods/c5878bc8-5cd8-4a31-8c7e-530618fb4848/volumes" Nov 23 07:09:35 crc kubenswrapper[4906]: I1123 07:09:35.378713 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5770119-9b13-474e-b564-a183d2dd9bb5" path="/var/lib/kubelet/pods/d5770119-9b13-474e-b564-a183d2dd9bb5/volumes" Nov 23 07:09:36 crc kubenswrapper[4906]: I1123 07:09:36.630971 4906 generic.go:334] "Generic (PLEG): container finished" podID="de58c3f3-2425-4137-8c3c-9871d1fbd9c1" containerID="cbd45e5087e3ff34aceaedab7642e3f533b4b897abad75afa1f69b5f5f11f37f" exitCode=0 Nov 23 07:09:36 crc kubenswrapper[4906]: I1123 07:09:36.631023 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qdmg" event={"ID":"de58c3f3-2425-4137-8c3c-9871d1fbd9c1","Type":"ContainerDied","Data":"cbd45e5087e3ff34aceaedab7642e3f533b4b897abad75afa1f69b5f5f11f37f"} Nov 23 07:09:39 crc kubenswrapper[4906]: I1123 07:09:39.495947 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:09:39 crc kubenswrapper[4906]: I1123 07:09:39.586732 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-hjdbc"] Nov 23 07:09:39 crc kubenswrapper[4906]: I1123 07:09:39.587532 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="dnsmasq-dns" containerID="cri-o://b4cc6a652fa7198326e843a437170ddeab2364ef675e70647d868d7c0942777a" gracePeriod=10 Nov 23 07:09:40 crc kubenswrapper[4906]: I1123 07:09:40.498119 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Nov 23 07:09:40 crc kubenswrapper[4906]: I1123 07:09:40.681451 4906 generic.go:334] "Generic (PLEG): container finished" podID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerID="b4cc6a652fa7198326e843a437170ddeab2364ef675e70647d868d7c0942777a" exitCode=0 Nov 23 07:09:40 crc kubenswrapper[4906]: I1123 07:09:40.681602 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" event={"ID":"d0e1a933-a98c-4401-99cd-5c1ad2873911","Type":"ContainerDied","Data":"b4cc6a652fa7198326e843a437170ddeab2364ef675e70647d868d7c0942777a"} Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.584475 4906 scope.go:117] "RemoveContainer" containerID="8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.679513 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.690594 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-config-data\") pod \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.690741 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-credential-keys\") pod \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.690939 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-scripts\") pod \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.690994 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-combined-ca-bundle\") pod \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.691075 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8sd2\" (UniqueName: \"kubernetes.io/projected/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-kube-api-access-j8sd2\") pod \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.691248 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-fernet-keys\") pod \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\" (UID: \"de58c3f3-2425-4137-8c3c-9871d1fbd9c1\") " Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.701163 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "de58c3f3-2425-4137-8c3c-9871d1fbd9c1" (UID: "de58c3f3-2425-4137-8c3c-9871d1fbd9c1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.702346 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qdmg" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.703084 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qdmg" event={"ID":"de58c3f3-2425-4137-8c3c-9871d1fbd9c1","Type":"ContainerDied","Data":"55b22750cb735ba0af2f22a508eacdaaed269239b61910400c6ad3c92e46ea23"} Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.703135 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55b22750cb735ba0af2f22a508eacdaaed269239b61910400c6ad3c92e46ea23" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.703186 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "de58c3f3-2425-4137-8c3c-9871d1fbd9c1" (UID: "de58c3f3-2425-4137-8c3c-9871d1fbd9c1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.720204 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-kube-api-access-j8sd2" (OuterVolumeSpecName: "kube-api-access-j8sd2") pod "de58c3f3-2425-4137-8c3c-9871d1fbd9c1" (UID: "de58c3f3-2425-4137-8c3c-9871d1fbd9c1"). InnerVolumeSpecName "kube-api-access-j8sd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.726970 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-scripts" (OuterVolumeSpecName: "scripts") pod "de58c3f3-2425-4137-8c3c-9871d1fbd9c1" (UID: "de58c3f3-2425-4137-8c3c-9871d1fbd9c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.731219 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de58c3f3-2425-4137-8c3c-9871d1fbd9c1" (UID: "de58c3f3-2425-4137-8c3c-9871d1fbd9c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.752423 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-config-data" (OuterVolumeSpecName: "config-data") pod "de58c3f3-2425-4137-8c3c-9871d1fbd9c1" (UID: "de58c3f3-2425-4137-8c3c-9871d1fbd9c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.793325 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.793358 4906 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.793368 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.793376 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.793385 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8sd2\" (UniqueName: \"kubernetes.io/projected/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-kube-api-access-j8sd2\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:41 crc kubenswrapper[4906]: I1123 07:09:41.793394 4906 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/de58c3f3-2425-4137-8c3c-9871d1fbd9c1-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.772142 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9qdmg"] Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.780650 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9qdmg"] Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.887238 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-78bh5"] Nov 23 07:09:42 crc kubenswrapper[4906]: E1123 07:09:42.887656 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de58c3f3-2425-4137-8c3c-9871d1fbd9c1" containerName="keystone-bootstrap" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.887692 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="de58c3f3-2425-4137-8c3c-9871d1fbd9c1" containerName="keystone-bootstrap" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.887856 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="de58c3f3-2425-4137-8c3c-9871d1fbd9c1" containerName="keystone-bootstrap" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.888494 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.891329 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.891432 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.891590 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nk8ll" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.895461 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.895543 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.902641 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-78bh5"] Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.946807 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-scripts\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.946860 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-config-data\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.946878 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-credential-keys\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.946899 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-fernet-keys\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.946991 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fzkt\" (UniqueName: \"kubernetes.io/projected/d86996ce-011e-4d9b-85c0-dae8ba298edd-kube-api-access-5fzkt\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:42 crc kubenswrapper[4906]: I1123 07:09:42.947022 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-combined-ca-bundle\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.048653 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fzkt\" (UniqueName: \"kubernetes.io/projected/d86996ce-011e-4d9b-85c0-dae8ba298edd-kube-api-access-5fzkt\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.048743 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-combined-ca-bundle\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.048832 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-scripts\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.048853 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-config-data\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.048869 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-credential-keys\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.048891 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-fernet-keys\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.055385 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-credential-keys\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.055416 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-config-data\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.055967 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-fernet-keys\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.057005 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-scripts\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.059445 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-combined-ca-bundle\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.082470 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fzkt\" (UniqueName: \"kubernetes.io/projected/d86996ce-011e-4d9b-85c0-dae8ba298edd-kube-api-access-5fzkt\") pod \"keystone-bootstrap-78bh5\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.289324 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:43 crc kubenswrapper[4906]: I1123 07:09:43.378493 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de58c3f3-2425-4137-8c3c-9871d1fbd9c1" path="/var/lib/kubelet/pods/de58c3f3-2425-4137-8c3c-9871d1fbd9c1/volumes" Nov 23 07:09:50 crc kubenswrapper[4906]: I1123 07:09:50.497315 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Nov 23 07:09:50 crc kubenswrapper[4906]: E1123 07:09:50.891234 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140" Nov 23 07:09:50 crc kubenswrapper[4906]: E1123 07:09:50.891614 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nddh5d6h557h647h5bbhbch56h5ch655h5cdh698h657h58ch5b4h56dh56h558hfh5b9h58bh589h598h647h59bh5d9h674h55bhcbhf9h66fh54fh698q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z8w7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(865cb0c4-2164-4c89-9153-20f19e15db0c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:09:50 crc kubenswrapper[4906]: I1123 07:09:50.945634 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:09:50 crc kubenswrapper[4906]: I1123 07:09:50.945882 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.042442 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.119185 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-sb\") pod \"d0e1a933-a98c-4401-99cd-5c1ad2873911\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.119254 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45fz2\" (UniqueName: \"kubernetes.io/projected/d0e1a933-a98c-4401-99cd-5c1ad2873911-kube-api-access-45fz2\") pod \"d0e1a933-a98c-4401-99cd-5c1ad2873911\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.119313 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-swift-storage-0\") pod \"d0e1a933-a98c-4401-99cd-5c1ad2873911\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.119387 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-svc\") pod \"d0e1a933-a98c-4401-99cd-5c1ad2873911\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.119405 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-nb\") pod \"d0e1a933-a98c-4401-99cd-5c1ad2873911\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.119513 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-config\") pod \"d0e1a933-a98c-4401-99cd-5c1ad2873911\" (UID: \"d0e1a933-a98c-4401-99cd-5c1ad2873911\") " Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.145605 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0e1a933-a98c-4401-99cd-5c1ad2873911-kube-api-access-45fz2" (OuterVolumeSpecName: "kube-api-access-45fz2") pod "d0e1a933-a98c-4401-99cd-5c1ad2873911" (UID: "d0e1a933-a98c-4401-99cd-5c1ad2873911"). InnerVolumeSpecName "kube-api-access-45fz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.176377 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d0e1a933-a98c-4401-99cd-5c1ad2873911" (UID: "d0e1a933-a98c-4401-99cd-5c1ad2873911"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.178222 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0e1a933-a98c-4401-99cd-5c1ad2873911" (UID: "d0e1a933-a98c-4401-99cd-5c1ad2873911"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.180802 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d0e1a933-a98c-4401-99cd-5c1ad2873911" (UID: "d0e1a933-a98c-4401-99cd-5c1ad2873911"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.183098 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d0e1a933-a98c-4401-99cd-5c1ad2873911" (UID: "d0e1a933-a98c-4401-99cd-5c1ad2873911"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.187499 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-config" (OuterVolumeSpecName: "config") pod "d0e1a933-a98c-4401-99cd-5c1ad2873911" (UID: "d0e1a933-a98c-4401-99cd-5c1ad2873911"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.221557 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.221593 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.221608 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45fz2\" (UniqueName: \"kubernetes.io/projected/d0e1a933-a98c-4401-99cd-5c1ad2873911-kube-api-access-45fz2\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.221618 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.221627 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.221636 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0e1a933-a98c-4401-99cd-5c1ad2873911-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.504767 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.816347 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" event={"ID":"d0e1a933-a98c-4401-99cd-5c1ad2873911","Type":"ContainerDied","Data":"0d2a0502aefd5cedbddfdf97893e5e72684697537c09fda9a9ef69d5adf71c62"} Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.816439 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.843154 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-hjdbc"] Nov 23 07:09:51 crc kubenswrapper[4906]: I1123 07:09:51.851187 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-hjdbc"] Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.131446 4906 scope.go:117] "RemoveContainer" containerID="60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742" Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.132964 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742\": container with ID starting with 60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742 not found: ID does not exist" containerID="60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.133042 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742"} err="failed to get container status \"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742\": rpc error: code = NotFound desc = could not find container \"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742\": container with ID starting with 60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742 not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.133094 4906 scope.go:117] "RemoveContainer" containerID="8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc" Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.133652 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc\": container with ID starting with 8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc not found: ID does not exist" containerID="8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.133753 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc"} err="failed to get container status \"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc\": rpc error: code = NotFound desc = could not find container \"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc\": container with ID starting with 8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.133808 4906 scope.go:117] "RemoveContainer" containerID="60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.134514 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742"} err="failed to get container status \"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742\": rpc error: code = NotFound desc = could not find container \"60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742\": container with ID starting with 60774a6301dcd8510ea127e7eb94f2c70c7db73d1e691259aa38e40f212b9742 not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.134568 4906 scope.go:117] "RemoveContainer" containerID="8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.135008 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc"} err="failed to get container status \"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc\": rpc error: code = NotFound desc = could not find container \"8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc\": container with ID starting with 8401a5d20bbf6712a1051e2af5a73510a853bbf1bc70d7515cb91046067c19fc not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.135054 4906 scope.go:117] "RemoveContainer" containerID="f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367" Nov 23 07:09:52 crc kubenswrapper[4906]: W1123 07:09:52.152951 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6209532_4ce6_482f_9c69_021b4ec0b682.slice/crio-28c48540a505a37e47f0494c78d96f8d1c56492f04c126483447b18c686c854b WatchSource:0}: Error finding container 28c48540a505a37e47f0494c78d96f8d1c56492f04c126483447b18c686c854b: Status 404 returned error can't find the container with id 28c48540a505a37e47f0494c78d96f8d1c56492f04c126483447b18c686c854b Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.166979 4906 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879" Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.167313 4906 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zjxdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-6p4bd_openstack(d2d3d804-9bab-439f-94c6-23e1546aaad7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.168865 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-6p4bd" podUID="d2d3d804-9bab-439f-94c6-23e1546aaad7" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.356463 4906 scope.go:117] "RemoveContainer" containerID="0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.460416 4906 scope.go:117] "RemoveContainer" containerID="f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367" Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.466424 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367\": container with ID starting with f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367 not found: ID does not exist" containerID="f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.466512 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367"} err="failed to get container status \"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367\": rpc error: code = NotFound desc = could not find container \"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367\": container with ID starting with f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367 not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.466565 4906 scope.go:117] "RemoveContainer" containerID="0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e" Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.467187 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e\": container with ID starting with 0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e not found: ID does not exist" containerID="0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.467221 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e"} err="failed to get container status \"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e\": rpc error: code = NotFound desc = could not find container \"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e\": container with ID starting with 0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.467240 4906 scope.go:117] "RemoveContainer" containerID="f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.467669 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367"} err="failed to get container status \"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367\": rpc error: code = NotFound desc = could not find container \"f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367\": container with ID starting with f27b11b4ac2a7f7a5761bae4ec163c4504c386760e1409f9fd302b56cd10e367 not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.467743 4906 scope.go:117] "RemoveContainer" containerID="0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.468208 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e"} err="failed to get container status \"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e\": rpc error: code = NotFound desc = could not find container \"0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e\": container with ID starting with 0b066f0dbb9f2c35ffee9c9f24b9cc32a56da4b5044f4ebdc0ccef204945000e not found: ID does not exist" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.468241 4906 scope.go:117] "RemoveContainer" containerID="b4cc6a652fa7198326e843a437170ddeab2364ef675e70647d868d7c0942777a" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.501224 4906 scope.go:117] "RemoveContainer" containerID="2aa7e0410d9580fdc0f37c3e6733c159984772ff90d8d592f658e9ad0a53886b" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.699604 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-78bh5"] Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.724955 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.856582 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6wvm5" event={"ID":"8f23c4a7-b2b9-4ef9-876b-5dd95f486333","Type":"ContainerStarted","Data":"b7828b55bc0c0687cf95e500a55863b74ca5ad9cf605f0ed87a24da3adab43cb"} Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.860908 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e6209532-4ce6-482f-9c69-021b4ec0b682","Type":"ContainerStarted","Data":"28c48540a505a37e47f0494c78d96f8d1c56492f04c126483447b18c686c854b"} Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.869432 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9l4vd" event={"ID":"5584696b-3782-4db6-997e-9f691f70d05a","Type":"ContainerStarted","Data":"64f8b0bc40e85813f6b53ebb1774deed7359fad207a70108c893d8f0a3d3af91"} Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.884984 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6wvm5" podStartSLOduration=4.241733112 podStartE2EDuration="24.884967703s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="2025-11-23 07:09:30.271901861 +0000 UTC m=+1185.785293164" lastFinishedPulling="2025-11-23 07:09:50.915136412 +0000 UTC m=+1206.428527755" observedRunningTime="2025-11-23 07:09:52.880630179 +0000 UTC m=+1208.394021482" watchObservedRunningTime="2025-11-23 07:09:52.884967703 +0000 UTC m=+1208.398359006" Nov 23 07:09:52 crc kubenswrapper[4906]: E1123 07:09:52.886778 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879\\\"\"" pod="openstack/cinder-db-sync-6p4bd" podUID="d2d3d804-9bab-439f-94c6-23e1546aaad7" Nov 23 07:09:52 crc kubenswrapper[4906]: I1123 07:09:52.901448 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9l4vd" podStartSLOduration=3.064048405 podStartE2EDuration="24.901429183s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="2025-11-23 07:09:30.342776992 +0000 UTC m=+1185.856168285" lastFinishedPulling="2025-11-23 07:09:52.18015776 +0000 UTC m=+1207.693549063" observedRunningTime="2025-11-23 07:09:52.897364656 +0000 UTC m=+1208.410755959" watchObservedRunningTime="2025-11-23 07:09:52.901429183 +0000 UTC m=+1208.414820486" Nov 23 07:09:53 crc kubenswrapper[4906]: W1123 07:09:53.050959 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e594b90_2851_4374_8913_08103bb5065e.slice/crio-883e3792178ea678c841a89a6558d5631e7685262cc51a58ddfe8c25d79faad4 WatchSource:0}: Error finding container 883e3792178ea678c841a89a6558d5631e7685262cc51a58ddfe8c25d79faad4: Status 404 returned error can't find the container with id 883e3792178ea678c841a89a6558d5631e7685262cc51a58ddfe8c25d79faad4 Nov 23 07:09:53 crc kubenswrapper[4906]: W1123 07:09:53.053410 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd86996ce_011e_4d9b_85c0_dae8ba298edd.slice/crio-4094e5a9346e975c612e603465967a1ea2d1b29f5b09279f505829f436329f39 WatchSource:0}: Error finding container 4094e5a9346e975c612e603465967a1ea2d1b29f5b09279f505829f436329f39: Status 404 returned error can't find the container with id 4094e5a9346e975c612e603465967a1ea2d1b29f5b09279f505829f436329f39 Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.070178 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.372390 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" path="/var/lib/kubelet/pods/d0e1a933-a98c-4401-99cd-5c1ad2873911/volumes" Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.932401 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e6209532-4ce6-482f-9c69-021b4ec0b682","Type":"ContainerStarted","Data":"dae906aff5a2080ffb6904c4f1c75020ff73527712718bb4c1180d689e9bae6d"} Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.934223 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e6209532-4ce6-482f-9c69-021b4ec0b682","Type":"ContainerStarted","Data":"a7efd54a95aafa3d3cc917715d3b8d6474962d24f6d6688799722b7635d6e399"} Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.938988 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerStarted","Data":"f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e"} Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.946435 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9e594b90-2851-4374-8913-08103bb5065e","Type":"ContainerStarted","Data":"2b9c7e90e2950b3fcc525bdf5a7dbb42733038fec6d65bd737ad2199a9071b46"} Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.946485 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9e594b90-2851-4374-8913-08103bb5065e","Type":"ContainerStarted","Data":"883e3792178ea678c841a89a6558d5631e7685262cc51a58ddfe8c25d79faad4"} Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.951037 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-78bh5" event={"ID":"d86996ce-011e-4d9b-85c0-dae8ba298edd","Type":"ContainerStarted","Data":"a132f35e3274982037e61dc2cbbb3c263680061b23f75f55481fc4121df6fdc6"} Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.951117 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-78bh5" event={"ID":"d86996ce-011e-4d9b-85c0-dae8ba298edd","Type":"ContainerStarted","Data":"4094e5a9346e975c612e603465967a1ea2d1b29f5b09279f505829f436329f39"} Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.981143 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=19.981102492 podStartE2EDuration="19.981102492s" podCreationTimestamp="2025-11-23 07:09:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:53.954638082 +0000 UTC m=+1209.468029395" watchObservedRunningTime="2025-11-23 07:09:53.981102492 +0000 UTC m=+1209.494493795" Nov 23 07:09:53 crc kubenswrapper[4906]: I1123 07:09:53.983712 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-78bh5" podStartSLOduration=11.98370297 podStartE2EDuration="11.98370297s" podCreationTimestamp="2025-11-23 07:09:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:53.977605822 +0000 UTC m=+1209.490997125" watchObservedRunningTime="2025-11-23 07:09:53.98370297 +0000 UTC m=+1209.497094273" Nov 23 07:09:54 crc kubenswrapper[4906]: I1123 07:09:54.997831 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 07:09:54 crc kubenswrapper[4906]: I1123 07:09:54.998234 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 07:09:55 crc kubenswrapper[4906]: I1123 07:09:55.051909 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 07:09:55 crc kubenswrapper[4906]: I1123 07:09:55.054961 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 07:09:55 crc kubenswrapper[4906]: I1123 07:09:55.499867 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6856c564b9-hjdbc" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Nov 23 07:09:55 crc kubenswrapper[4906]: I1123 07:09:55.971115 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9e594b90-2851-4374-8913-08103bb5065e","Type":"ContainerStarted","Data":"20b05ca7f58640a43e4cde5f3c87862d66748a5431dc1655c260c2a29bf3dae2"} Nov 23 07:09:55 crc kubenswrapper[4906]: I1123 07:09:55.971739 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 07:09:55 crc kubenswrapper[4906]: I1123 07:09:55.971797 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 07:09:56 crc kubenswrapper[4906]: I1123 07:09:55.994751 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.994727779 podStartE2EDuration="21.994727779s" podCreationTimestamp="2025-11-23 07:09:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:09:55.991600936 +0000 UTC m=+1211.504992249" watchObservedRunningTime="2025-11-23 07:09:55.994727779 +0000 UTC m=+1211.508119092" Nov 23 07:09:57 crc kubenswrapper[4906]: I1123 07:09:57.995613 4906 generic.go:334] "Generic (PLEG): container finished" podID="d86996ce-011e-4d9b-85c0-dae8ba298edd" containerID="a132f35e3274982037e61dc2cbbb3c263680061b23f75f55481fc4121df6fdc6" exitCode=0 Nov 23 07:09:57 crc kubenswrapper[4906]: I1123 07:09:57.995801 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-78bh5" event={"ID":"d86996ce-011e-4d9b-85c0-dae8ba298edd","Type":"ContainerDied","Data":"a132f35e3274982037e61dc2cbbb3c263680061b23f75f55481fc4121df6fdc6"} Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.013449 4906 generic.go:334] "Generic (PLEG): container finished" podID="8f23c4a7-b2b9-4ef9-876b-5dd95f486333" containerID="b7828b55bc0c0687cf95e500a55863b74ca5ad9cf605f0ed87a24da3adab43cb" exitCode=0 Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.013653 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6wvm5" event={"ID":"8f23c4a7-b2b9-4ef9-876b-5dd95f486333","Type":"ContainerDied","Data":"b7828b55bc0c0687cf95e500a55863b74ca5ad9cf605f0ed87a24da3adab43cb"} Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.508056 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.634883 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-config-data\") pod \"d86996ce-011e-4d9b-85c0-dae8ba298edd\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.635015 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-fernet-keys\") pod \"d86996ce-011e-4d9b-85c0-dae8ba298edd\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.635065 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fzkt\" (UniqueName: \"kubernetes.io/projected/d86996ce-011e-4d9b-85c0-dae8ba298edd-kube-api-access-5fzkt\") pod \"d86996ce-011e-4d9b-85c0-dae8ba298edd\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.635128 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-combined-ca-bundle\") pod \"d86996ce-011e-4d9b-85c0-dae8ba298edd\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.635275 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-credential-keys\") pod \"d86996ce-011e-4d9b-85c0-dae8ba298edd\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.635335 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-scripts\") pod \"d86996ce-011e-4d9b-85c0-dae8ba298edd\" (UID: \"d86996ce-011e-4d9b-85c0-dae8ba298edd\") " Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.641668 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d86996ce-011e-4d9b-85c0-dae8ba298edd" (UID: "d86996ce-011e-4d9b-85c0-dae8ba298edd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.642875 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d86996ce-011e-4d9b-85c0-dae8ba298edd" (UID: "d86996ce-011e-4d9b-85c0-dae8ba298edd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.644149 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-scripts" (OuterVolumeSpecName: "scripts") pod "d86996ce-011e-4d9b-85c0-dae8ba298edd" (UID: "d86996ce-011e-4d9b-85c0-dae8ba298edd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.645632 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d86996ce-011e-4d9b-85c0-dae8ba298edd-kube-api-access-5fzkt" (OuterVolumeSpecName: "kube-api-access-5fzkt") pod "d86996ce-011e-4d9b-85c0-dae8ba298edd" (UID: "d86996ce-011e-4d9b-85c0-dae8ba298edd"). InnerVolumeSpecName "kube-api-access-5fzkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.671582 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d86996ce-011e-4d9b-85c0-dae8ba298edd" (UID: "d86996ce-011e-4d9b-85c0-dae8ba298edd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.678791 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-config-data" (OuterVolumeSpecName: "config-data") pod "d86996ce-011e-4d9b-85c0-dae8ba298edd" (UID: "d86996ce-011e-4d9b-85c0-dae8ba298edd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.739079 4906 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.739123 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.739134 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.739146 4906 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.739159 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fzkt\" (UniqueName: \"kubernetes.io/projected/d86996ce-011e-4d9b-85c0-dae8ba298edd-kube-api-access-5fzkt\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:59 crc kubenswrapper[4906]: I1123 07:09:59.739176 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86996ce-011e-4d9b-85c0-dae8ba298edd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.023084 4906 generic.go:334] "Generic (PLEG): container finished" podID="5584696b-3782-4db6-997e-9f691f70d05a" containerID="64f8b0bc40e85813f6b53ebb1774deed7359fad207a70108c893d8f0a3d3af91" exitCode=0 Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.023259 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9l4vd" event={"ID":"5584696b-3782-4db6-997e-9f691f70d05a","Type":"ContainerDied","Data":"64f8b0bc40e85813f6b53ebb1774deed7359fad207a70108c893d8f0a3d3af91"} Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.027343 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-78bh5" event={"ID":"d86996ce-011e-4d9b-85c0-dae8ba298edd","Type":"ContainerDied","Data":"4094e5a9346e975c612e603465967a1ea2d1b29f5b09279f505829f436329f39"} Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.027368 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4094e5a9346e975c612e603465967a1ea2d1b29f5b09279f505829f436329f39" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.027623 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-78bh5" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.034754 4906 generic.go:334] "Generic (PLEG): container finished" podID="0cc86f52-0139-430f-9803-06b714d9fc7e" containerID="739fbb582901e5e7181fb6829f18d4295a29898fbc690f867445b5383dea4f42" exitCode=0 Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.034802 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ltbgj" event={"ID":"0cc86f52-0139-430f-9803-06b714d9fc7e","Type":"ContainerDied","Data":"739fbb582901e5e7181fb6829f18d4295a29898fbc690f867445b5383dea4f42"} Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.037522 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerStarted","Data":"cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192"} Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.155569 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-87fd667c9-tzc42"] Nov 23 07:10:00 crc kubenswrapper[4906]: E1123 07:10:00.156795 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d86996ce-011e-4d9b-85c0-dae8ba298edd" containerName="keystone-bootstrap" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.156863 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d86996ce-011e-4d9b-85c0-dae8ba298edd" containerName="keystone-bootstrap" Nov 23 07:10:00 crc kubenswrapper[4906]: E1123 07:10:00.157088 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="init" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.157140 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="init" Nov 23 07:10:00 crc kubenswrapper[4906]: E1123 07:10:00.157196 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="dnsmasq-dns" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.157242 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="dnsmasq-dns" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.157498 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0e1a933-a98c-4401-99cd-5c1ad2873911" containerName="dnsmasq-dns" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.157555 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d86996ce-011e-4d9b-85c0-dae8ba298edd" containerName="keystone-bootstrap" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.158478 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.164989 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.165118 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.165434 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.165577 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.165849 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-nk8ll" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.173954 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-87fd667c9-tzc42"] Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.176148 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.253816 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-config-data\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.253955 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-credential-keys\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.254016 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-fernet-keys\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.254040 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9gbw\" (UniqueName: \"kubernetes.io/projected/2e303c56-48b8-4176-99f6-fd0aeb26cf94-kube-api-access-d9gbw\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.254104 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-scripts\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.254260 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-internal-tls-certs\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.254348 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-public-tls-certs\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.254437 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-combined-ca-bundle\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358217 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-scripts\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358280 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-internal-tls-certs\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358320 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-public-tls-certs\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358358 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-combined-ca-bundle\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358381 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-config-data\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358421 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-credential-keys\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358444 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-fernet-keys\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.358461 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9gbw\" (UniqueName: \"kubernetes.io/projected/2e303c56-48b8-4176-99f6-fd0aeb26cf94-kube-api-access-d9gbw\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.366100 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-public-tls-certs\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.366124 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-internal-tls-certs\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.366852 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-combined-ca-bundle\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.367603 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-credential-keys\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.367809 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-scripts\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.368861 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-config-data\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.372148 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-fernet-keys\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.375108 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9gbw\" (UniqueName: \"kubernetes.io/projected/2e303c56-48b8-4176-99f6-fd0aeb26cf94-kube-api-access-d9gbw\") pod \"keystone-87fd667c9-tzc42\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.448457 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6wvm5" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.493230 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.562544 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-scripts\") pod \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.562743 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-logs\") pod \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.563741 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-combined-ca-bundle\") pod \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.563732 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-logs" (OuterVolumeSpecName: "logs") pod "8f23c4a7-b2b9-4ef9-876b-5dd95f486333" (UID: "8f23c4a7-b2b9-4ef9-876b-5dd95f486333"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.563887 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-config-data\") pod \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.564039 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tgkp\" (UniqueName: \"kubernetes.io/projected/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-kube-api-access-7tgkp\") pod \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\" (UID: \"8f23c4a7-b2b9-4ef9-876b-5dd95f486333\") " Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.565166 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.569867 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-scripts" (OuterVolumeSpecName: "scripts") pod "8f23c4a7-b2b9-4ef9-876b-5dd95f486333" (UID: "8f23c4a7-b2b9-4ef9-876b-5dd95f486333"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.571317 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-kube-api-access-7tgkp" (OuterVolumeSpecName: "kube-api-access-7tgkp") pod "8f23c4a7-b2b9-4ef9-876b-5dd95f486333" (UID: "8f23c4a7-b2b9-4ef9-876b-5dd95f486333"). InnerVolumeSpecName "kube-api-access-7tgkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.615396 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f23c4a7-b2b9-4ef9-876b-5dd95f486333" (UID: "8f23c4a7-b2b9-4ef9-876b-5dd95f486333"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.618408 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-config-data" (OuterVolumeSpecName: "config-data") pod "8f23c4a7-b2b9-4ef9-876b-5dd95f486333" (UID: "8f23c4a7-b2b9-4ef9-876b-5dd95f486333"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.667811 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tgkp\" (UniqueName: \"kubernetes.io/projected/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-kube-api-access-7tgkp\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.667869 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.667884 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.667894 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f23c4a7-b2b9-4ef9-876b-5dd95f486333-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:00 crc kubenswrapper[4906]: I1123 07:10:00.951368 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-87fd667c9-tzc42"] Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.059875 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6wvm5" event={"ID":"8f23c4a7-b2b9-4ef9-876b-5dd95f486333","Type":"ContainerDied","Data":"e65b6796fc5a43c9b616dbcb5c30b54532e6c9c0e24b12e56d3ccbd0811e8029"} Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.060378 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e65b6796fc5a43c9b616dbcb5c30b54532e6c9c0e24b12e56d3ccbd0811e8029" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.059021 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6wvm5" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.068001 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87fd667c9-tzc42" event={"ID":"2e303c56-48b8-4176-99f6-fd0aeb26cf94","Type":"ContainerStarted","Data":"c8ec7974ff0eb8d271d91f3ef8c4b9b542caa3a6f488bea691468c7520ddd20f"} Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.270032 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ffb7cbc86-lz7bc"] Nov 23 07:10:01 crc kubenswrapper[4906]: E1123 07:10:01.271712 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f23c4a7-b2b9-4ef9-876b-5dd95f486333" containerName="placement-db-sync" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.271729 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f23c4a7-b2b9-4ef9-876b-5dd95f486333" containerName="placement-db-sync" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.271905 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f23c4a7-b2b9-4ef9-876b-5dd95f486333" containerName="placement-db-sync" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.278810 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ffb7cbc86-lz7bc"] Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.280370 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.286943 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.287210 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-59ssp" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.287375 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.287512 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.287533 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.381605 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73149d5c-c16f-4b83-b855-135fc1aab9fe-logs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.382976 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-internal-tls-certs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.383074 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-combined-ca-bundle\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.383313 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-config-data\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.383360 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw487\" (UniqueName: \"kubernetes.io/projected/73149d5c-c16f-4b83-b855-135fc1aab9fe-kube-api-access-vw487\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.383463 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-public-tls-certs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.383618 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-scripts\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.439219 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.485777 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-config-data\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.485825 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw487\" (UniqueName: \"kubernetes.io/projected/73149d5c-c16f-4b83-b855-135fc1aab9fe-kube-api-access-vw487\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.485908 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-public-tls-certs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.485962 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-scripts\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.486038 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73149d5c-c16f-4b83-b855-135fc1aab9fe-logs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.486080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-internal-tls-certs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.486134 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-combined-ca-bundle\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.487797 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73149d5c-c16f-4b83-b855-135fc1aab9fe-logs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.493465 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-config-data\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.494077 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-internal-tls-certs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.495003 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-scripts\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.495635 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-combined-ca-bundle\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.504943 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-public-tls-certs\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.508965 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw487\" (UniqueName: \"kubernetes.io/projected/73149d5c-c16f-4b83-b855-135fc1aab9fe-kube-api-access-vw487\") pod \"placement-ffb7cbc86-lz7bc\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.588611 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-combined-ca-bundle\") pod \"5584696b-3782-4db6-997e-9f691f70d05a\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.588672 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-db-sync-config-data\") pod \"5584696b-3782-4db6-997e-9f691f70d05a\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.588734 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2smc\" (UniqueName: \"kubernetes.io/projected/5584696b-3782-4db6-997e-9f691f70d05a-kube-api-access-d2smc\") pod \"5584696b-3782-4db6-997e-9f691f70d05a\" (UID: \"5584696b-3782-4db6-997e-9f691f70d05a\") " Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.592854 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5584696b-3782-4db6-997e-9f691f70d05a" (UID: "5584696b-3782-4db6-997e-9f691f70d05a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.593540 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5584696b-3782-4db6-997e-9f691f70d05a-kube-api-access-d2smc" (OuterVolumeSpecName: "kube-api-access-d2smc") pod "5584696b-3782-4db6-997e-9f691f70d05a" (UID: "5584696b-3782-4db6-997e-9f691f70d05a"). InnerVolumeSpecName "kube-api-access-d2smc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.608198 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.615367 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5584696b-3782-4db6-997e-9f691f70d05a" (UID: "5584696b-3782-4db6-997e-9f691f70d05a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.617014 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.702296 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-combined-ca-bundle\") pod \"0cc86f52-0139-430f-9803-06b714d9fc7e\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.702513 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-config\") pod \"0cc86f52-0139-430f-9803-06b714d9fc7e\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.702573 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrsg9\" (UniqueName: \"kubernetes.io/projected/0cc86f52-0139-430f-9803-06b714d9fc7e-kube-api-access-lrsg9\") pod \"0cc86f52-0139-430f-9803-06b714d9fc7e\" (UID: \"0cc86f52-0139-430f-9803-06b714d9fc7e\") " Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.704604 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.704627 4906 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5584696b-3782-4db6-997e-9f691f70d05a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.704639 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2smc\" (UniqueName: \"kubernetes.io/projected/5584696b-3782-4db6-997e-9f691f70d05a-kube-api-access-d2smc\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.706622 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc86f52-0139-430f-9803-06b714d9fc7e-kube-api-access-lrsg9" (OuterVolumeSpecName: "kube-api-access-lrsg9") pod "0cc86f52-0139-430f-9803-06b714d9fc7e" (UID: "0cc86f52-0139-430f-9803-06b714d9fc7e"). InnerVolumeSpecName "kube-api-access-lrsg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.728646 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-config" (OuterVolumeSpecName: "config") pod "0cc86f52-0139-430f-9803-06b714d9fc7e" (UID: "0cc86f52-0139-430f-9803-06b714d9fc7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.735911 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cc86f52-0139-430f-9803-06b714d9fc7e" (UID: "0cc86f52-0139-430f-9803-06b714d9fc7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.816846 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.817355 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cc86f52-0139-430f-9803-06b714d9fc7e-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:01 crc kubenswrapper[4906]: I1123 07:10:01.817366 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrsg9\" (UniqueName: \"kubernetes.io/projected/0cc86f52-0139-430f-9803-06b714d9fc7e-kube-api-access-lrsg9\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.103115 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9l4vd" event={"ID":"5584696b-3782-4db6-997e-9f691f70d05a","Type":"ContainerDied","Data":"6e20a48620a788b724cb505b39fd8e4ce789fc5289f259790e1fa5ead16977ff"} Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.103168 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e20a48620a788b724cb505b39fd8e4ce789fc5289f259790e1fa5ead16977ff" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.103311 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9l4vd" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.110301 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87fd667c9-tzc42" event={"ID":"2e303c56-48b8-4176-99f6-fd0aeb26cf94","Type":"ContainerStarted","Data":"4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb"} Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.110382 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.117441 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ltbgj" event={"ID":"0cc86f52-0139-430f-9803-06b714d9fc7e","Type":"ContainerDied","Data":"ca0dfaa3b3b681f02dcd40ec304d45c72edc9225c7d05e2761efe5001a5e6a60"} Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.117489 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca0dfaa3b3b681f02dcd40ec304d45c72edc9225c7d05e2761efe5001a5e6a60" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.117518 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ltbgj" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.163081 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-87fd667c9-tzc42" podStartSLOduration=2.163028151 podStartE2EDuration="2.163028151s" podCreationTimestamp="2025-11-23 07:10:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:02.145421761 +0000 UTC m=+1217.658813064" watchObservedRunningTime="2025-11-23 07:10:02.163028151 +0000 UTC m=+1217.676419454" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.199987 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ffb7cbc86-lz7bc"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.290025 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-pr9gz"] Nov 23 07:10:02 crc kubenswrapper[4906]: E1123 07:10:02.298032 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc86f52-0139-430f-9803-06b714d9fc7e" containerName="neutron-db-sync" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.298058 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc86f52-0139-430f-9803-06b714d9fc7e" containerName="neutron-db-sync" Nov 23 07:10:02 crc kubenswrapper[4906]: E1123 07:10:02.298079 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5584696b-3782-4db6-997e-9f691f70d05a" containerName="barbican-db-sync" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.298086 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5584696b-3782-4db6-997e-9f691f70d05a" containerName="barbican-db-sync" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.298267 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5584696b-3782-4db6-997e-9f691f70d05a" containerName="barbican-db-sync" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.298288 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc86f52-0139-430f-9803-06b714d9fc7e" containerName="neutron-db-sync" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.299384 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.332618 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.332702 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-svc\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.332735 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.333059 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.333298 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7b7g\" (UniqueName: \"kubernetes.io/projected/73e048ea-b143-411b-bdea-bd6ebe093c87-kube-api-access-v7b7g\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.333353 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-config\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.363845 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-pr9gz"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.376231 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6f9d47b994-ptdjw"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.379299 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.384337 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.384598 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.386814 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b2ntq" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.427056 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6f9d47b994-ptdjw"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.436848 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-svc\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.436922 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.436987 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce87b9c-add6-4744-8463-0d268867b9de-logs\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437035 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437128 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7b7g\" (UniqueName: \"kubernetes.io/projected/73e048ea-b143-411b-bdea-bd6ebe093c87-kube-api-access-v7b7g\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437160 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-combined-ca-bundle\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437187 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-config\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437207 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437266 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zld8t\" (UniqueName: \"kubernetes.io/projected/0ce87b9c-add6-4744-8463-0d268867b9de-kube-api-access-zld8t\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437298 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data-custom\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.437372 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.438134 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-svc\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.446222 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.456110 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.456707 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.459489 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-config\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.460769 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5584fdbb5-9b5k8"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.462465 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.468658 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7b7g\" (UniqueName: \"kubernetes.io/projected/73e048ea-b143-411b-bdea-bd6ebe093c87-kube-api-access-v7b7g\") pod \"dnsmasq-dns-6c654c9745-pr9gz\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.478357 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.489467 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5584fdbb5-9b5k8"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.529847 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-56d695cc56-fkbp4"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.531452 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.534320 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.535213 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.535379 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4555g" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.535467 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.552942 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-combined-ca-bundle\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.553096 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.553175 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zld8t\" (UniqueName: \"kubernetes.io/projected/0ce87b9c-add6-4744-8463-0d268867b9de-kube-api-access-zld8t\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.553272 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data-custom\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.553427 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce87b9c-add6-4744-8463-0d268867b9de-logs\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.555263 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce87b9c-add6-4744-8463-0d268867b9de-logs\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.564102 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56d695cc56-fkbp4"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.564341 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-combined-ca-bundle\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.577021 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.577508 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data-custom\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.586539 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zld8t\" (UniqueName: \"kubernetes.io/projected/0ce87b9c-add6-4744-8463-0d268867b9de-kube-api-access-zld8t\") pod \"barbican-keystone-listener-6f9d47b994-ptdjw\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.634841 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-pr9gz"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.635525 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.654975 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-ovndb-tls-certs\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655058 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655077 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-combined-ca-bundle\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655101 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-httpd-config\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655141 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf8fg\" (UniqueName: \"kubernetes.io/projected/2d3c7e43-fe35-428c-99ae-35245b5d62b6-kube-api-access-vf8fg\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655165 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3c7e43-fe35-428c-99ae-35245b5d62b6-logs\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655210 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-combined-ca-bundle\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655246 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-config\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655269 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data-custom\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.655297 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbqwv\" (UniqueName: \"kubernetes.io/projected/00e11c87-d705-4b7a-8771-13d3117e67e7-kube-api-access-wbqwv\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.704633 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-z9rz8"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.706105 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.707479 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.722101 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-68644c84-m657h"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.725092 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.736277 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-z9rz8"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.746984 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.755593 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-68644c84-m657h"] Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.759541 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.761846 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-combined-ca-bundle\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.762738 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-httpd-config\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.763318 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf8fg\" (UniqueName: \"kubernetes.io/projected/2d3c7e43-fe35-428c-99ae-35245b5d62b6-kube-api-access-vf8fg\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.763928 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3c7e43-fe35-428c-99ae-35245b5d62b6-logs\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.766037 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3c7e43-fe35-428c-99ae-35245b5d62b6-logs\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.767575 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-combined-ca-bundle\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.771222 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-config\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.769340 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.771999 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data-custom\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.772302 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbqwv\" (UniqueName: \"kubernetes.io/projected/00e11c87-d705-4b7a-8771-13d3117e67e7-kube-api-access-wbqwv\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.772332 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-combined-ca-bundle\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.772485 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-ovndb-tls-certs\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.773583 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-httpd-config\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.776508 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-config\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.779628 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data-custom\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.781496 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-combined-ca-bundle\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.783705 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-ovndb-tls-certs\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.794721 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf8fg\" (UniqueName: \"kubernetes.io/projected/2d3c7e43-fe35-428c-99ae-35245b5d62b6-kube-api-access-vf8fg\") pod \"barbican-worker-5584fdbb5-9b5k8\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.799057 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbqwv\" (UniqueName: \"kubernetes.io/projected/00e11c87-d705-4b7a-8771-13d3117e67e7-kube-api-access-wbqwv\") pod \"neutron-56d695cc56-fkbp4\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.813267 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.867024 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875294 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875348 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875412 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875451 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zpd5\" (UniqueName: \"kubernetes.io/projected/495c2958-59c5-4eb7-a9ee-299e675cf8ff-kube-api-access-2zpd5\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875552 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data-custom\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875581 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875639 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-config\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875663 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s655\" (UniqueName: \"kubernetes.io/projected/ccf9323c-773e-49e4-8788-0cc4fddcbb09-kube-api-access-5s655\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875725 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccf9323c-773e-49e4-8788-0cc4fddcbb09-logs\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875758 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.875784 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-combined-ca-bundle\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.979752 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data-custom\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.979820 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.979869 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-config\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.979904 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s655\" (UniqueName: \"kubernetes.io/projected/ccf9323c-773e-49e4-8788-0cc4fddcbb09-kube-api-access-5s655\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.979934 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccf9323c-773e-49e4-8788-0cc4fddcbb09-logs\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.979961 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.979999 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-combined-ca-bundle\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.980043 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.980072 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.980111 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.980133 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zpd5\" (UniqueName: \"kubernetes.io/projected/495c2958-59c5-4eb7-a9ee-299e675cf8ff-kube-api-access-2zpd5\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.980795 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccf9323c-773e-49e4-8788-0cc4fddcbb09-logs\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.982011 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-config\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.983588 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.985221 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.985793 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.987301 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.987856 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data-custom\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.988475 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-combined-ca-bundle\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.989421 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:02 crc kubenswrapper[4906]: I1123 07:10:02.997308 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zpd5\" (UniqueName: \"kubernetes.io/projected/495c2958-59c5-4eb7-a9ee-299e675cf8ff-kube-api-access-2zpd5\") pod \"dnsmasq-dns-5cc67f459c-z9rz8\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:02.999900 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s655\" (UniqueName: \"kubernetes.io/projected/ccf9323c-773e-49e4-8788-0cc4fddcbb09-kube-api-access-5s655\") pod \"barbican-api-68644c84-m657h\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.041289 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-pr9gz"] Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.050407 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.074849 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.136053 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" event={"ID":"73e048ea-b143-411b-bdea-bd6ebe093c87","Type":"ContainerStarted","Data":"42cda90c7b18bd3482dfe98a713ce801b405ae18b5563c624378a3303b677712"} Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.145602 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ffb7cbc86-lz7bc" event={"ID":"73149d5c-c16f-4b83-b855-135fc1aab9fe","Type":"ContainerStarted","Data":"5bd4727bfb09b7d4d7d82d1343a0b814629a6bdd84d7c99f84522b703b964091"} Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.418950 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6f9d47b994-ptdjw"] Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.528596 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5584fdbb5-9b5k8"] Nov 23 07:10:03 crc kubenswrapper[4906]: W1123 07:10:03.533890 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d3c7e43_fe35_428c_99ae_35245b5d62b6.slice/crio-aa72bd392bf223fd68c4c4e2eb0a7967c0b90a9916d600182f377941466826f0 WatchSource:0}: Error finding container aa72bd392bf223fd68c4c4e2eb0a7967c0b90a9916d600182f377941466826f0: Status 404 returned error can't find the container with id aa72bd392bf223fd68c4c4e2eb0a7967c0b90a9916d600182f377941466826f0 Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.678636 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-68644c84-m657h"] Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.692030 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-z9rz8"] Nov 23 07:10:03 crc kubenswrapper[4906]: W1123 07:10:03.705050 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccf9323c_773e_49e4_8788_0cc4fddcbb09.slice/crio-2838b650edab37ea36c04e8bea9f5159a43698a221607fc325214688b1e6fc31 WatchSource:0}: Error finding container 2838b650edab37ea36c04e8bea9f5159a43698a221607fc325214688b1e6fc31: Status 404 returned error can't find the container with id 2838b650edab37ea36c04e8bea9f5159a43698a221607fc325214688b1e6fc31 Nov 23 07:10:03 crc kubenswrapper[4906]: W1123 07:10:03.708444 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod495c2958_59c5_4eb7_a9ee_299e675cf8ff.slice/crio-cd04ecab952d4e13899cfb8e0f1d5e8081956cb0882d142a26d46f88012d1a43 WatchSource:0}: Error finding container cd04ecab952d4e13899cfb8e0f1d5e8081956cb0882d142a26d46f88012d1a43: Status 404 returned error can't find the container with id cd04ecab952d4e13899cfb8e0f1d5e8081956cb0882d142a26d46f88012d1a43 Nov 23 07:10:03 crc kubenswrapper[4906]: I1123 07:10:03.875448 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56d695cc56-fkbp4"] Nov 23 07:10:03 crc kubenswrapper[4906]: W1123 07:10:03.895065 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00e11c87_d705_4b7a_8771_13d3117e67e7.slice/crio-82959acd674edb822ef56fc5944fb4da0381b1425ca7deece625633c7690ccb8 WatchSource:0}: Error finding container 82959acd674edb822ef56fc5944fb4da0381b1425ca7deece625633c7690ccb8: Status 404 returned error can't find the container with id 82959acd674edb822ef56fc5944fb4da0381b1425ca7deece625633c7690ccb8 Nov 23 07:10:04 crc kubenswrapper[4906]: I1123 07:10:04.162324 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" event={"ID":"495c2958-59c5-4eb7-a9ee-299e675cf8ff","Type":"ContainerStarted","Data":"cd04ecab952d4e13899cfb8e0f1d5e8081956cb0882d142a26d46f88012d1a43"} Nov 23 07:10:04 crc kubenswrapper[4906]: I1123 07:10:04.168276 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" event={"ID":"73e048ea-b143-411b-bdea-bd6ebe093c87","Type":"ContainerStarted","Data":"cf446cb3f5b10a6a88d8e28317dc2585487bd722492562cb66b85393f4c20009"} Nov 23 07:10:04 crc kubenswrapper[4906]: I1123 07:10:04.170784 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d695cc56-fkbp4" event={"ID":"00e11c87-d705-4b7a-8771-13d3117e67e7","Type":"ContainerStarted","Data":"82959acd674edb822ef56fc5944fb4da0381b1425ca7deece625633c7690ccb8"} Nov 23 07:10:04 crc kubenswrapper[4906]: I1123 07:10:04.172287 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5584fdbb5-9b5k8" event={"ID":"2d3c7e43-fe35-428c-99ae-35245b5d62b6","Type":"ContainerStarted","Data":"aa72bd392bf223fd68c4c4e2eb0a7967c0b90a9916d600182f377941466826f0"} Nov 23 07:10:04 crc kubenswrapper[4906]: I1123 07:10:04.173798 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" event={"ID":"0ce87b9c-add6-4744-8463-0d268867b9de","Type":"ContainerStarted","Data":"a5997aa130cada46fbd35cadfd9b32721747769ca4e08c725a23e4faa8f1983c"} Nov 23 07:10:04 crc kubenswrapper[4906]: I1123 07:10:04.178794 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68644c84-m657h" event={"ID":"ccf9323c-773e-49e4-8788-0cc4fddcbb09","Type":"ContainerStarted","Data":"2838b650edab37ea36c04e8bea9f5159a43698a221607fc325214688b1e6fc31"} Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.166490 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7f84fc787f-zskfn"] Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.172043 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.174203 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.174588 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.179033 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f84fc787f-zskfn"] Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.198423 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.199530 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.199578 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.204886 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.205406 4906 generic.go:334] "Generic (PLEG): container finished" podID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerID="b084e4cd79cd99c0619e4f93a8893304d5cf1f0f0eefbcdca66d25decbb61cda" exitCode=0 Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.205480 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" event={"ID":"495c2958-59c5-4eb7-a9ee-299e675cf8ff","Type":"ContainerDied","Data":"b084e4cd79cd99c0619e4f93a8893304d5cf1f0f0eefbcdca66d25decbb61cda"} Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.210396 4906 generic.go:334] "Generic (PLEG): container finished" podID="73e048ea-b143-411b-bdea-bd6ebe093c87" containerID="cf446cb3f5b10a6a88d8e28317dc2585487bd722492562cb66b85393f4c20009" exitCode=0 Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.210468 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" event={"ID":"73e048ea-b143-411b-bdea-bd6ebe093c87","Type":"ContainerDied","Data":"cf446cb3f5b10a6a88d8e28317dc2585487bd722492562cb66b85393f4c20009"} Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.227892 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d695cc56-fkbp4" event={"ID":"00e11c87-d705-4b7a-8771-13d3117e67e7","Type":"ContainerStarted","Data":"c0578e8d6b5b396123727d6c012bf4d48a3e9c86ec3bea27282318f3e04bfffa"} Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.235127 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-config\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.235209 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plqzn\" (UniqueName: \"kubernetes.io/projected/e84af371-9517-42aa-943b-a4847f4f8577-kube-api-access-plqzn\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.235356 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-ovndb-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.235394 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-internal-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.235511 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-public-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.235615 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-httpd-config\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.235672 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-combined-ca-bundle\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.250262 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68644c84-m657h" event={"ID":"ccf9323c-773e-49e4-8788-0cc4fddcbb09","Type":"ContainerStarted","Data":"c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb"} Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.283920 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.338586 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-httpd-config\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.339027 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-combined-ca-bundle\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.339077 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-config\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.339115 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plqzn\" (UniqueName: \"kubernetes.io/projected/e84af371-9517-42aa-943b-a4847f4f8577-kube-api-access-plqzn\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.339167 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-ovndb-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.339202 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-internal-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.339249 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-public-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.348826 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-public-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.351605 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-httpd-config\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.352611 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-config\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.359462 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-internal-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.359525 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-combined-ca-bundle\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.360665 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-ovndb-tls-certs\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.375486 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plqzn\" (UniqueName: \"kubernetes.io/projected/e84af371-9517-42aa-943b-a4847f4f8577-kube-api-access-plqzn\") pod \"neutron-7f84fc787f-zskfn\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.455048 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:05 crc kubenswrapper[4906]: I1123 07:10:05.503954 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.224848 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f84fc787f-zskfn"] Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.278846 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" event={"ID":"495c2958-59c5-4eb7-a9ee-299e675cf8ff","Type":"ContainerStarted","Data":"e9a97e802ec3d2f28e7ab2b9a9e4811701dbf5d52cbc5746484f45eee01458d6"} Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.280202 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.293063 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d695cc56-fkbp4" event={"ID":"00e11c87-d705-4b7a-8771-13d3117e67e7","Type":"ContainerStarted","Data":"ad06d562f80a95bcf1c5e5404e0c38aa2d345c32e7bce7a62c5cf872ff3fe9f4"} Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.293256 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.299569 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ffb7cbc86-lz7bc" event={"ID":"73149d5c-c16f-4b83-b855-135fc1aab9fe","Type":"ContainerStarted","Data":"a3add9b8cbef96dced17c8263087b9f52097f904de7cf1efc850f3e41c400acf"} Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.307881 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68644c84-m657h" event={"ID":"ccf9323c-773e-49e4-8788-0cc4fddcbb09","Type":"ContainerStarted","Data":"bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094"} Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.308410 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.308591 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.315805 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" podStartSLOduration=4.315721697 podStartE2EDuration="4.315721697s" podCreationTimestamp="2025-11-23 07:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:06.305055518 +0000 UTC m=+1221.818446831" watchObservedRunningTime="2025-11-23 07:10:06.315721697 +0000 UTC m=+1221.829113000" Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.355081 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-56d695cc56-fkbp4" podStartSLOduration=4.355048214 podStartE2EDuration="4.355048214s" podCreationTimestamp="2025-11-23 07:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:06.324713561 +0000 UTC m=+1221.838104864" watchObservedRunningTime="2025-11-23 07:10:06.355048214 +0000 UTC m=+1221.868439517" Nov 23 07:10:06 crc kubenswrapper[4906]: I1123 07:10:06.361043 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-68644c84-m657h" podStartSLOduration=4.36102813 podStartE2EDuration="4.36102813s" podCreationTimestamp="2025-11-23 07:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:06.352034285 +0000 UTC m=+1221.865425588" watchObservedRunningTime="2025-11-23 07:10:06.36102813 +0000 UTC m=+1221.874419433" Nov 23 07:10:07 crc kubenswrapper[4906]: I1123 07:10:07.759348 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 07:10:07 crc kubenswrapper[4906]: I1123 07:10:07.795868 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 07:10:08 crc kubenswrapper[4906]: I1123 07:10:08.621992 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:08 crc kubenswrapper[4906]: I1123 07:10:08.624384 4906 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 07:10:08 crc kubenswrapper[4906]: I1123 07:10:08.828989 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.440620 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-bdb7fc4cb-h962s"] Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.445448 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.450713 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bdb7fc4cb-h962s"] Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.452434 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.457860 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.548929 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data-custom\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.549060 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w7n6\" (UniqueName: \"kubernetes.io/projected/4aedece6-346c-4003-8d00-95509a35f23b-kube-api-access-9w7n6\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.549120 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.549149 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-internal-tls-certs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.549186 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-public-tls-certs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.549231 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aedece6-346c-4003-8d00-95509a35f23b-logs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.549249 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-combined-ca-bundle\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.651304 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data-custom\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.651382 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w7n6\" (UniqueName: \"kubernetes.io/projected/4aedece6-346c-4003-8d00-95509a35f23b-kube-api-access-9w7n6\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.651456 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.651489 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-internal-tls-certs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.651534 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-public-tls-certs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.651617 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aedece6-346c-4003-8d00-95509a35f23b-logs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.651635 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-combined-ca-bundle\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.652996 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aedece6-346c-4003-8d00-95509a35f23b-logs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.660443 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.666988 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-combined-ca-bundle\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.667191 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-public-tls-certs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.667657 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-internal-tls-certs\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.669306 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data-custom\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.669847 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w7n6\" (UniqueName: \"kubernetes.io/projected/4aedece6-346c-4003-8d00-95509a35f23b-kube-api-access-9w7n6\") pod \"barbican-api-bdb7fc4cb-h962s\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:09 crc kubenswrapper[4906]: I1123 07:10:09.787997 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:11 crc kubenswrapper[4906]: W1123 07:10:11.102881 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode84af371_9517_42aa_943b_a4847f4f8577.slice/crio-41bed39932780cf18202c0714f56879367261dc38f9854bf05fdceb5faee7e35 WatchSource:0}: Error finding container 41bed39932780cf18202c0714f56879367261dc38f9854bf05fdceb5faee7e35: Status 404 returned error can't find the container with id 41bed39932780cf18202c0714f56879367261dc38f9854bf05fdceb5faee7e35 Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.208703 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.291152 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-svc\") pod \"73e048ea-b143-411b-bdea-bd6ebe093c87\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.291223 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-nb\") pod \"73e048ea-b143-411b-bdea-bd6ebe093c87\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.292398 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7b7g\" (UniqueName: \"kubernetes.io/projected/73e048ea-b143-411b-bdea-bd6ebe093c87-kube-api-access-v7b7g\") pod \"73e048ea-b143-411b-bdea-bd6ebe093c87\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.292503 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-sb\") pod \"73e048ea-b143-411b-bdea-bd6ebe093c87\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.292566 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-config\") pod \"73e048ea-b143-411b-bdea-bd6ebe093c87\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.292595 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-swift-storage-0\") pod \"73e048ea-b143-411b-bdea-bd6ebe093c87\" (UID: \"73e048ea-b143-411b-bdea-bd6ebe093c87\") " Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.299971 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e048ea-b143-411b-bdea-bd6ebe093c87-kube-api-access-v7b7g" (OuterVolumeSpecName: "kube-api-access-v7b7g") pod "73e048ea-b143-411b-bdea-bd6ebe093c87" (UID: "73e048ea-b143-411b-bdea-bd6ebe093c87"). InnerVolumeSpecName "kube-api-access-v7b7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.324467 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "73e048ea-b143-411b-bdea-bd6ebe093c87" (UID: "73e048ea-b143-411b-bdea-bd6ebe093c87"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.324646 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73e048ea-b143-411b-bdea-bd6ebe093c87" (UID: "73e048ea-b143-411b-bdea-bd6ebe093c87"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.331042 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-config" (OuterVolumeSpecName: "config") pod "73e048ea-b143-411b-bdea-bd6ebe093c87" (UID: "73e048ea-b143-411b-bdea-bd6ebe093c87"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.333545 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73e048ea-b143-411b-bdea-bd6ebe093c87" (UID: "73e048ea-b143-411b-bdea-bd6ebe093c87"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.334613 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73e048ea-b143-411b-bdea-bd6ebe093c87" (UID: "73e048ea-b143-411b-bdea-bd6ebe093c87"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.371972 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f84fc787f-zskfn" event={"ID":"e84af371-9517-42aa-943b-a4847f4f8577","Type":"ContainerStarted","Data":"41bed39932780cf18202c0714f56879367261dc38f9854bf05fdceb5faee7e35"} Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.374019 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" event={"ID":"73e048ea-b143-411b-bdea-bd6ebe093c87","Type":"ContainerDied","Data":"42cda90c7b18bd3482dfe98a713ce801b405ae18b5563c624378a3303b677712"} Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.374119 4906 scope.go:117] "RemoveContainer" containerID="cf446cb3f5b10a6a88d8e28317dc2585487bd722492562cb66b85393f4c20009" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.374334 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-pr9gz" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.396028 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.396063 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.396079 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.396092 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.396104 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e048ea-b143-411b-bdea-bd6ebe093c87-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.396115 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7b7g\" (UniqueName: \"kubernetes.io/projected/73e048ea-b143-411b-bdea-bd6ebe093c87-kube-api-access-v7b7g\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.441467 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-pr9gz"] Nov 23 07:10:11 crc kubenswrapper[4906]: I1123 07:10:11.452006 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-pr9gz"] Nov 23 07:10:13 crc kubenswrapper[4906]: I1123 07:10:13.054922 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:13 crc kubenswrapper[4906]: I1123 07:10:13.149227 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-595vk"] Nov 23 07:10:13 crc kubenswrapper[4906]: I1123 07:10:13.149881 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerName="dnsmasq-dns" containerID="cri-o://bdc38c6e92d3fc410dab382cf993c78b39f4c8731349dc9479864fc24e34c033" gracePeriod=10 Nov 23 07:10:13 crc kubenswrapper[4906]: I1123 07:10:13.383042 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e048ea-b143-411b-bdea-bd6ebe093c87" path="/var/lib/kubelet/pods/73e048ea-b143-411b-bdea-bd6ebe093c87/volumes" Nov 23 07:10:13 crc kubenswrapper[4906]: I1123 07:10:13.403592 4906 generic.go:334] "Generic (PLEG): container finished" podID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerID="bdc38c6e92d3fc410dab382cf993c78b39f4c8731349dc9479864fc24e34c033" exitCode=0 Nov 23 07:10:13 crc kubenswrapper[4906]: I1123 07:10:13.403637 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" event={"ID":"1e49ca90-ea62-4f3b-88de-bc876ef28d65","Type":"ContainerDied","Data":"bdc38c6e92d3fc410dab382cf993c78b39f4c8731349dc9479864fc24e34c033"} Nov 23 07:10:14 crc kubenswrapper[4906]: I1123 07:10:14.666330 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:14 crc kubenswrapper[4906]: I1123 07:10:14.731789 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.458203 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.481692 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.482439 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" event={"ID":"1e49ca90-ea62-4f3b-88de-bc876ef28d65","Type":"ContainerDied","Data":"fc438511ee7772182cc30c2f3d1b3f0e0e3b6a741fa3abf59d3cdc70ef8cdeb8"} Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.482489 4906 scope.go:117] "RemoveContainer" containerID="bdc38c6e92d3fc410dab382cf993c78b39f4c8731349dc9479864fc24e34c033" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.642799 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-swift-storage-0\") pod \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.642859 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-config\") pod \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.642994 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-nb\") pod \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.643095 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf5w5\" (UniqueName: \"kubernetes.io/projected/1e49ca90-ea62-4f3b-88de-bc876ef28d65-kube-api-access-xf5w5\") pod \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.643119 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-svc\") pod \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.643146 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-sb\") pod \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\" (UID: \"1e49ca90-ea62-4f3b-88de-bc876ef28d65\") " Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.650671 4906 scope.go:117] "RemoveContainer" containerID="b1307d8fa8d02b6b3c9334b2d3d293699ee2cb419fb6b42499ad3b0351985caa" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.673299 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e49ca90-ea62-4f3b-88de-bc876ef28d65-kube-api-access-xf5w5" (OuterVolumeSpecName: "kube-api-access-xf5w5") pod "1e49ca90-ea62-4f3b-88de-bc876ef28d65" (UID: "1e49ca90-ea62-4f3b-88de-bc876ef28d65"). InnerVolumeSpecName "kube-api-access-xf5w5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.747793 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf5w5\" (UniqueName: \"kubernetes.io/projected/1e49ca90-ea62-4f3b-88de-bc876ef28d65-kube-api-access-xf5w5\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.768990 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bdb7fc4cb-h962s"] Nov 23 07:10:15 crc kubenswrapper[4906]: W1123 07:10:15.781366 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aedece6_346c_4003_8d00_95509a35f23b.slice/crio-84b441d187fa8cdc082ceec636414ade0519a551c14dc2185e814ec8ae3a3647 WatchSource:0}: Error finding container 84b441d187fa8cdc082ceec636414ade0519a551c14dc2185e814ec8ae3a3647: Status 404 returned error can't find the container with id 84b441d187fa8cdc082ceec636414ade0519a551c14dc2185e814ec8ae3a3647 Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.870498 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1e49ca90-ea62-4f3b-88de-bc876ef28d65" (UID: "1e49ca90-ea62-4f3b-88de-bc876ef28d65"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.875045 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1e49ca90-ea62-4f3b-88de-bc876ef28d65" (UID: "1e49ca90-ea62-4f3b-88de-bc876ef28d65"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.892574 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-config" (OuterVolumeSpecName: "config") pod "1e49ca90-ea62-4f3b-88de-bc876ef28d65" (UID: "1e49ca90-ea62-4f3b-88de-bc876ef28d65"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.895952 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1e49ca90-ea62-4f3b-88de-bc876ef28d65" (UID: "1e49ca90-ea62-4f3b-88de-bc876ef28d65"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.897470 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1e49ca90-ea62-4f3b-88de-bc876ef28d65" (UID: "1e49ca90-ea62-4f3b-88de-bc876ef28d65"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.952427 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.952665 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.952768 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.952856 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:15 crc kubenswrapper[4906]: I1123 07:10:15.952927 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e49ca90-ea62-4f3b-88de-bc876ef28d65-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:16 crc kubenswrapper[4906]: E1123 07:10:16.055309 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.155854 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-595vk"] Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.160605 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-595vk"] Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.500329 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6p4bd" event={"ID":"d2d3d804-9bab-439f-94c6-23e1546aaad7","Type":"ContainerStarted","Data":"1f3a9d97d10c50b09e77dcba87dffe97001871bb27409ed0d6f948816f5a32ec"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.524554 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" event={"ID":"0ce87b9c-add6-4744-8463-0d268867b9de","Type":"ContainerStarted","Data":"9acde65807d055f979c4c0e765b70fcc9e174d4e93f5057b281b3cd782bbd787"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.525135 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" event={"ID":"0ce87b9c-add6-4744-8463-0d268867b9de","Type":"ContainerStarted","Data":"e5685a80b861544fdb6e63f53e8fc51d7652169e4cb7aa78555e20b2db5bcf62"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.543083 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f84fc787f-zskfn" event={"ID":"e84af371-9517-42aa-943b-a4847f4f8577","Type":"ContainerStarted","Data":"5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.543138 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f84fc787f-zskfn" event={"ID":"e84af371-9517-42aa-943b-a4847f4f8577","Type":"ContainerStarted","Data":"dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.544178 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.550110 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5584fdbb5-9b5k8" event={"ID":"2d3c7e43-fe35-428c-99ae-35245b5d62b6","Type":"ContainerStarted","Data":"9488749cc23de30cd7d4a1998d3ecd858dad70cda1dd749aa047283ebdd924bf"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.553873 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-6p4bd" podStartSLOduration=3.287230146 podStartE2EDuration="48.553840668s" podCreationTimestamp="2025-11-23 07:09:28 +0000 UTC" firstStartedPulling="2025-11-23 07:09:30.031213696 +0000 UTC m=+1185.544604999" lastFinishedPulling="2025-11-23 07:10:15.297824218 +0000 UTC m=+1230.811215521" observedRunningTime="2025-11-23 07:10:16.524400617 +0000 UTC m=+1232.037791930" watchObservedRunningTime="2025-11-23 07:10:16.553840668 +0000 UTC m=+1232.067231981" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.556243 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ffb7cbc86-lz7bc" event={"ID":"73149d5c-c16f-4b83-b855-135fc1aab9fe","Type":"ContainerStarted","Data":"e5fc6b77a408f81d5b91fd45ba27278acb2c5d247b602dbc2475870c5ba45b81"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.556586 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.556616 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.574881 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bdb7fc4cb-h962s" event={"ID":"4aedece6-346c-4003-8d00-95509a35f23b","Type":"ContainerStarted","Data":"78e2be376a6ac18267e9fa24359c2479618c77eeaf00a6b4b20186ab50f161d6"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.574935 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bdb7fc4cb-h962s" event={"ID":"4aedece6-346c-4003-8d00-95509a35f23b","Type":"ContainerStarted","Data":"84b441d187fa8cdc082ceec636414ade0519a551c14dc2185e814ec8ae3a3647"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.576812 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" podStartSLOduration=2.702124965 podStartE2EDuration="14.57680095s" podCreationTimestamp="2025-11-23 07:10:02 +0000 UTC" firstStartedPulling="2025-11-23 07:10:03.423106992 +0000 UTC m=+1218.936498305" lastFinishedPulling="2025-11-23 07:10:15.297782987 +0000 UTC m=+1230.811174290" observedRunningTime="2025-11-23 07:10:16.55161905 +0000 UTC m=+1232.065010343" watchObservedRunningTime="2025-11-23 07:10:16.57680095 +0000 UTC m=+1232.090192253" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.579312 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5584fdbb5-9b5k8" podStartSLOduration=2.812709231 podStartE2EDuration="14.579305385s" podCreationTimestamp="2025-11-23 07:10:02 +0000 UTC" firstStartedPulling="2025-11-23 07:10:03.537805366 +0000 UTC m=+1219.051196659" lastFinishedPulling="2025-11-23 07:10:15.30440152 +0000 UTC m=+1230.817792813" observedRunningTime="2025-11-23 07:10:16.572435765 +0000 UTC m=+1232.085827068" watchObservedRunningTime="2025-11-23 07:10:16.579305385 +0000 UTC m=+1232.092696678" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.586979 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerStarted","Data":"9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420"} Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.587100 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="ceilometer-notification-agent" containerID="cri-o://f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e" gracePeriod=30 Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.587186 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.587224 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="proxy-httpd" containerID="cri-o://9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420" gracePeriod=30 Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.587258 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="sg-core" containerID="cri-o://cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192" gracePeriod=30 Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.594461 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7f84fc787f-zskfn" podStartSLOduration=11.594445352 podStartE2EDuration="11.594445352s" podCreationTimestamp="2025-11-23 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:16.593392164 +0000 UTC m=+1232.106783467" watchObservedRunningTime="2025-11-23 07:10:16.594445352 +0000 UTC m=+1232.107836655" Nov 23 07:10:16 crc kubenswrapper[4906]: I1123 07:10:16.643398 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-ffb7cbc86-lz7bc" podStartSLOduration=15.643377414 podStartE2EDuration="15.643377414s" podCreationTimestamp="2025-11-23 07:10:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:16.642281815 +0000 UTC m=+1232.155673118" watchObservedRunningTime="2025-11-23 07:10:16.643377414 +0000 UTC m=+1232.156768717" Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.378730 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" path="/var/lib/kubelet/pods/1e49ca90-ea62-4f3b-88de-bc876ef28d65/volumes" Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.599153 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5584fdbb5-9b5k8" event={"ID":"2d3c7e43-fe35-428c-99ae-35245b5d62b6","Type":"ContainerStarted","Data":"04f1ef97bd7ad66e51909908566ff20e01fcea5f37847fa8e0ca7e76df6a1b60"} Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.601797 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bdb7fc4cb-h962s" event={"ID":"4aedece6-346c-4003-8d00-95509a35f23b","Type":"ContainerStarted","Data":"0dd9b19587494f220b9ce7cb80f137d8caf2f6b1707a1f5bae2b3bc3625d12af"} Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.602111 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.602192 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.606314 4906 generic.go:334] "Generic (PLEG): container finished" podID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerID="9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420" exitCode=0 Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.606361 4906 generic.go:334] "Generic (PLEG): container finished" podID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerID="cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192" exitCode=2 Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.606395 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerDied","Data":"9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420"} Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.606466 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerDied","Data":"cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192"} Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.633557 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-bdb7fc4cb-h962s" podStartSLOduration=8.633494608 podStartE2EDuration="8.633494608s" podCreationTimestamp="2025-11-23 07:10:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:17.61943741 +0000 UTC m=+1233.132828723" watchObservedRunningTime="2025-11-23 07:10:17.633494608 +0000 UTC m=+1233.146885941" Nov 23 07:10:17 crc kubenswrapper[4906]: I1123 07:10:17.982291 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:19 crc kubenswrapper[4906]: I1123 07:10:19.496357 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76c58b6d97-595vk" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.290005 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.471008 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-sg-core-conf-yaml\") pod \"865cb0c4-2164-4c89-9153-20f19e15db0c\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.471078 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-config-data\") pod \"865cb0c4-2164-4c89-9153-20f19e15db0c\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.471174 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-combined-ca-bundle\") pod \"865cb0c4-2164-4c89-9153-20f19e15db0c\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.471213 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-scripts\") pod \"865cb0c4-2164-4c89-9153-20f19e15db0c\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.471346 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8w7w\" (UniqueName: \"kubernetes.io/projected/865cb0c4-2164-4c89-9153-20f19e15db0c-kube-api-access-z8w7w\") pod \"865cb0c4-2164-4c89-9153-20f19e15db0c\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.471429 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-log-httpd\") pod \"865cb0c4-2164-4c89-9153-20f19e15db0c\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.471650 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-run-httpd\") pod \"865cb0c4-2164-4c89-9153-20f19e15db0c\" (UID: \"865cb0c4-2164-4c89-9153-20f19e15db0c\") " Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.472161 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "865cb0c4-2164-4c89-9153-20f19e15db0c" (UID: "865cb0c4-2164-4c89-9153-20f19e15db0c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.472284 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "865cb0c4-2164-4c89-9153-20f19e15db0c" (UID: "865cb0c4-2164-4c89-9153-20f19e15db0c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.473405 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.473453 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865cb0c4-2164-4c89-9153-20f19e15db0c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.479942 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/865cb0c4-2164-4c89-9153-20f19e15db0c-kube-api-access-z8w7w" (OuterVolumeSpecName: "kube-api-access-z8w7w") pod "865cb0c4-2164-4c89-9153-20f19e15db0c" (UID: "865cb0c4-2164-4c89-9153-20f19e15db0c"). InnerVolumeSpecName "kube-api-access-z8w7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.480059 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-scripts" (OuterVolumeSpecName: "scripts") pod "865cb0c4-2164-4c89-9153-20f19e15db0c" (UID: "865cb0c4-2164-4c89-9153-20f19e15db0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.524707 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "865cb0c4-2164-4c89-9153-20f19e15db0c" (UID: "865cb0c4-2164-4c89-9153-20f19e15db0c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.550627 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "865cb0c4-2164-4c89-9153-20f19e15db0c" (UID: "865cb0c4-2164-4c89-9153-20f19e15db0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.575277 4906 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.575341 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.575369 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.575390 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8w7w\" (UniqueName: \"kubernetes.io/projected/865cb0c4-2164-4c89-9153-20f19e15db0c-kube-api-access-z8w7w\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.596094 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-config-data" (OuterVolumeSpecName: "config-data") pod "865cb0c4-2164-4c89-9153-20f19e15db0c" (UID: "865cb0c4-2164-4c89-9153-20f19e15db0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.645253 4906 generic.go:334] "Generic (PLEG): container finished" podID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerID="f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e" exitCode=0 Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.645315 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerDied","Data":"f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e"} Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.645386 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865cb0c4-2164-4c89-9153-20f19e15db0c","Type":"ContainerDied","Data":"4503eede6f143cfc69698a84c0e9db60a5906fa8a4e8c8427e2d1a838c44421e"} Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.645413 4906 scope.go:117] "RemoveContainer" containerID="9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.645832 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.671772 4906 scope.go:117] "RemoveContainer" containerID="cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.677465 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865cb0c4-2164-4c89-9153-20f19e15db0c-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.718801 4906 scope.go:117] "RemoveContainer" containerID="f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.741701 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.762746 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.772461 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.773030 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e048ea-b143-411b-bdea-bd6ebe093c87" containerName="init" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773054 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e048ea-b143-411b-bdea-bd6ebe093c87" containerName="init" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.773101 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerName="init" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773111 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerName="init" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.773127 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="sg-core" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773136 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="sg-core" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.773161 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="proxy-httpd" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773170 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="proxy-httpd" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.773190 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="ceilometer-notification-agent" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773200 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="ceilometer-notification-agent" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.773214 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerName="dnsmasq-dns" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773223 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerName="dnsmasq-dns" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773449 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e048ea-b143-411b-bdea-bd6ebe093c87" containerName="init" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773476 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="ceilometer-notification-agent" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773488 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="sg-core" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773501 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" containerName="proxy-httpd" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.773518 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e49ca90-ea62-4f3b-88de-bc876ef28d65" containerName="dnsmasq-dns" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.776324 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.780525 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.781861 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.782807 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.785582 4906 scope.go:117] "RemoveContainer" containerID="9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.789926 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420\": container with ID starting with 9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420 not found: ID does not exist" containerID="9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.789973 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420"} err="failed to get container status \"9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420\": rpc error: code = NotFound desc = could not find container \"9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420\": container with ID starting with 9a288c3067c7a38011c0a4885fd1a5f3a4c0836b486b06e9294a668ef24ea420 not found: ID does not exist" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.789993 4906 scope.go:117] "RemoveContainer" containerID="cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.790317 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192\": container with ID starting with cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192 not found: ID does not exist" containerID="cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.790332 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192"} err="failed to get container status \"cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192\": rpc error: code = NotFound desc = could not find container \"cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192\": container with ID starting with cab06dafe43fe60eca21137af26af60b453340177ceb83a83c9bcf1efdf08192 not found: ID does not exist" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.790343 4906 scope.go:117] "RemoveContainer" containerID="f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e" Nov 23 07:10:20 crc kubenswrapper[4906]: E1123 07:10:20.790543 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e\": container with ID starting with f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e not found: ID does not exist" containerID="f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.790562 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e"} err="failed to get container status \"f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e\": rpc error: code = NotFound desc = could not find container \"f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e\": container with ID starting with f5d335b6d820411c56026781f1536d5d9007d043710d1ae474dc38ed7370dc5e not found: ID does not exist" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.880552 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6ch\" (UniqueName: \"kubernetes.io/projected/4533768d-d850-457e-87cf-7adf0d594915-kube-api-access-zp6ch\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.880906 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-scripts\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.880942 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.880982 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-run-httpd\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.881001 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.881055 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-config-data\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.881098 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-log-httpd\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.945726 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.945829 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.983805 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6ch\" (UniqueName: \"kubernetes.io/projected/4533768d-d850-457e-87cf-7adf0d594915-kube-api-access-zp6ch\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.983889 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-scripts\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.983963 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.984064 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-run-httpd\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.984131 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.984264 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-config-data\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.984354 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-log-httpd\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.984748 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-run-httpd\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.984806 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-log-httpd\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.989054 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-config-data\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.989361 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.990116 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-scripts\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:20 crc kubenswrapper[4906]: I1123 07:10:20.990568 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:21 crc kubenswrapper[4906]: I1123 07:10:21.004207 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6ch\" (UniqueName: \"kubernetes.io/projected/4533768d-d850-457e-87cf-7adf0d594915-kube-api-access-zp6ch\") pod \"ceilometer-0\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " pod="openstack/ceilometer-0" Nov 23 07:10:21 crc kubenswrapper[4906]: I1123 07:10:21.098748 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:21 crc kubenswrapper[4906]: I1123 07:10:21.368496 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="865cb0c4-2164-4c89-9153-20f19e15db0c" path="/var/lib/kubelet/pods/865cb0c4-2164-4c89-9153-20f19e15db0c/volumes" Nov 23 07:10:21 crc kubenswrapper[4906]: I1123 07:10:21.551376 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:21 crc kubenswrapper[4906]: I1123 07:10:21.654878 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerStarted","Data":"557b365223e12f4583cbd2b9fcf0690a8da259f8e4e5d5b1f26befa836cbc34a"} Nov 23 07:10:21 crc kubenswrapper[4906]: I1123 07:10:21.657262 4906 generic.go:334] "Generic (PLEG): container finished" podID="d2d3d804-9bab-439f-94c6-23e1546aaad7" containerID="1f3a9d97d10c50b09e77dcba87dffe97001871bb27409ed0d6f948816f5a32ec" exitCode=0 Nov 23 07:10:21 crc kubenswrapper[4906]: I1123 07:10:21.657297 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6p4bd" event={"ID":"d2d3d804-9bab-439f-94c6-23e1546aaad7","Type":"ContainerDied","Data":"1f3a9d97d10c50b09e77dcba87dffe97001871bb27409ed0d6f948816f5a32ec"} Nov 23 07:10:22 crc kubenswrapper[4906]: I1123 07:10:22.720019 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerStarted","Data":"66dff64417cdf5fe7d1ce3e986378bd84df2c47d8565a82aeda164822b6780db"} Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.092663 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.236002 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-scripts\") pod \"d2d3d804-9bab-439f-94c6-23e1546aaad7\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.236171 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-db-sync-config-data\") pod \"d2d3d804-9bab-439f-94c6-23e1546aaad7\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.236266 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjxdv\" (UniqueName: \"kubernetes.io/projected/d2d3d804-9bab-439f-94c6-23e1546aaad7-kube-api-access-zjxdv\") pod \"d2d3d804-9bab-439f-94c6-23e1546aaad7\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.236347 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-config-data\") pod \"d2d3d804-9bab-439f-94c6-23e1546aaad7\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.236413 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-combined-ca-bundle\") pod \"d2d3d804-9bab-439f-94c6-23e1546aaad7\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.236463 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2d3d804-9bab-439f-94c6-23e1546aaad7-etc-machine-id\") pod \"d2d3d804-9bab-439f-94c6-23e1546aaad7\" (UID: \"d2d3d804-9bab-439f-94c6-23e1546aaad7\") " Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.236742 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2d3d804-9bab-439f-94c6-23e1546aaad7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d2d3d804-9bab-439f-94c6-23e1546aaad7" (UID: "d2d3d804-9bab-439f-94c6-23e1546aaad7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.237829 4906 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2d3d804-9bab-439f-94c6-23e1546aaad7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.241152 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-scripts" (OuterVolumeSpecName: "scripts") pod "d2d3d804-9bab-439f-94c6-23e1546aaad7" (UID: "d2d3d804-9bab-439f-94c6-23e1546aaad7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.242431 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d2d3d804-9bab-439f-94c6-23e1546aaad7" (UID: "d2d3d804-9bab-439f-94c6-23e1546aaad7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.244921 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d3d804-9bab-439f-94c6-23e1546aaad7-kube-api-access-zjxdv" (OuterVolumeSpecName: "kube-api-access-zjxdv") pod "d2d3d804-9bab-439f-94c6-23e1546aaad7" (UID: "d2d3d804-9bab-439f-94c6-23e1546aaad7"). InnerVolumeSpecName "kube-api-access-zjxdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.274829 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2d3d804-9bab-439f-94c6-23e1546aaad7" (UID: "d2d3d804-9bab-439f-94c6-23e1546aaad7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.291552 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-config-data" (OuterVolumeSpecName: "config-data") pod "d2d3d804-9bab-439f-94c6-23e1546aaad7" (UID: "d2d3d804-9bab-439f-94c6-23e1546aaad7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.339661 4906 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.339719 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjxdv\" (UniqueName: \"kubernetes.io/projected/d2d3d804-9bab-439f-94c6-23e1546aaad7-kube-api-access-zjxdv\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.339733 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.339750 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.339760 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d3d804-9bab-439f-94c6-23e1546aaad7-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.748135 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerStarted","Data":"f25969cc9818ffc774942bd7e1fc0db8a15e76f95b2f056249d5b1eec853d7e2"} Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.748184 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerStarted","Data":"122f5ddae4d56f6dd982766ee340c268fe423822feea55f993b8a2cb95d2b4a4"} Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.750172 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6p4bd" event={"ID":"d2d3d804-9bab-439f-94c6-23e1546aaad7","Type":"ContainerDied","Data":"4037c6ff51a31b197f817c2f158c2b2dc19ab3ae2763acb61c021a40cf648165"} Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.750193 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4037c6ff51a31b197f817c2f158c2b2dc19ab3ae2763acb61c021a40cf648165" Nov 23 07:10:23 crc kubenswrapper[4906]: I1123 07:10:23.750259 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6p4bd" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.091254 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:24 crc kubenswrapper[4906]: E1123 07:10:24.091926 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d3d804-9bab-439f-94c6-23e1546aaad7" containerName="cinder-db-sync" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.091946 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d3d804-9bab-439f-94c6-23e1546aaad7" containerName="cinder-db-sync" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.092155 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d3d804-9bab-439f-94c6-23e1546aaad7" containerName="cinder-db-sync" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.093098 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.096151 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-h6n2h" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.096559 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.096748 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.096888 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.109326 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.167587 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-797bbc649-d45mm"] Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.170707 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.190161 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-d45mm"] Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.268377 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.268449 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-657rw\" (UniqueName: \"kubernetes.io/projected/fee2c65e-0c82-4258-8263-3de044da4402-kube-api-access-657rw\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.268548 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06da91c4-1f07-43da-9462-1bc43fb8e255-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.268565 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.268832 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.268926 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-svc\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.268978 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.269102 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-config\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.269215 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.269290 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.269514 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jvpn\" (UniqueName: \"kubernetes.io/projected/06da91c4-1f07-43da-9462-1bc43fb8e255-kube-api-access-8jvpn\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.269552 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-scripts\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.324727 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.327200 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.329616 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.335969 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371749 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jvpn\" (UniqueName: \"kubernetes.io/projected/06da91c4-1f07-43da-9462-1bc43fb8e255-kube-api-access-8jvpn\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371799 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-scripts\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371840 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmjmq\" (UniqueName: \"kubernetes.io/projected/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-kube-api-access-zmjmq\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371860 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371888 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-657rw\" (UniqueName: \"kubernetes.io/projected/fee2c65e-0c82-4258-8263-3de044da4402-kube-api-access-657rw\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371917 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371953 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-logs\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371975 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06da91c4-1f07-43da-9462-1bc43fb8e255-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.371993 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372032 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data-custom\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372058 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372087 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-svc\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372109 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-scripts\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372124 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372141 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372162 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-config\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372186 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-etc-machine-id\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372203 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.372239 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.373210 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-svc\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.373793 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-config\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.374320 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.374367 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06da91c4-1f07-43da-9462-1bc43fb8e255-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.378089 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.378892 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.379050 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.379543 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.380488 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-scripts\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.381516 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.390530 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jvpn\" (UniqueName: \"kubernetes.io/projected/06da91c4-1f07-43da-9462-1bc43fb8e255-kube-api-access-8jvpn\") pod \"cinder-scheduler-0\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.393035 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-657rw\" (UniqueName: \"kubernetes.io/projected/fee2c65e-0c82-4258-8263-3de044da4402-kube-api-access-657rw\") pod \"dnsmasq-dns-797bbc649-d45mm\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.423285 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.473865 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmjmq\" (UniqueName: \"kubernetes.io/projected/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-kube-api-access-zmjmq\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.474719 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.475467 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-logs\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.475540 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data-custom\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.475705 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-scripts\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.475736 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.475791 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-etc-machine-id\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.475901 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-etc-machine-id\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.476787 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-logs\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.481458 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.484541 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data-custom\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.484606 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.484984 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-scripts\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.495220 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmjmq\" (UniqueName: \"kubernetes.io/projected/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-kube-api-access-zmjmq\") pod \"cinder-api-0\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.499818 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.644440 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.899100 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:24 crc kubenswrapper[4906]: W1123 07:10:24.943347 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06da91c4_1f07_43da_9462_1bc43fb8e255.slice/crio-726c22e1ea8e8427b1a09e37c542133371cf2bcb74e1104fb03062dcc561e09e WatchSource:0}: Error finding container 726c22e1ea8e8427b1a09e37c542133371cf2bcb74e1104fb03062dcc561e09e: Status 404 returned error can't find the container with id 726c22e1ea8e8427b1a09e37c542133371cf2bcb74e1104fb03062dcc561e09e Nov 23 07:10:24 crc kubenswrapper[4906]: I1123 07:10:24.966958 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:24 crc kubenswrapper[4906]: W1123 07:10:24.987750 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65ca8ef2_8cdd_4b3f_9705_03fdbbdcd218.slice/crio-36380df8e1a7af246a193138757eb8fd543ae3697add37578503345fc5642a1b WatchSource:0}: Error finding container 36380df8e1a7af246a193138757eb8fd543ae3697add37578503345fc5642a1b: Status 404 returned error can't find the container with id 36380df8e1a7af246a193138757eb8fd543ae3697add37578503345fc5642a1b Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.033577 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-d45mm"] Nov 23 07:10:25 crc kubenswrapper[4906]: W1123 07:10:25.063588 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfee2c65e_0c82_4258_8263_3de044da4402.slice/crio-a6074f7996d62eca3feb85117ab5e990011b33df5b80fc761b517f99e148b7e7 WatchSource:0}: Error finding container a6074f7996d62eca3feb85117ab5e990011b33df5b80fc761b517f99e148b7e7: Status 404 returned error can't find the container with id a6074f7996d62eca3feb85117ab5e990011b33df5b80fc761b517f99e148b7e7 Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.829384 4906 generic.go:334] "Generic (PLEG): container finished" podID="fee2c65e-0c82-4258-8263-3de044da4402" containerID="a6009fd7d930fa522902f4cc415e5ac1d70f0ced8343de6d0c4d564d6ae712c0" exitCode=0 Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.829468 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-d45mm" event={"ID":"fee2c65e-0c82-4258-8263-3de044da4402","Type":"ContainerDied","Data":"a6009fd7d930fa522902f4cc415e5ac1d70f0ced8343de6d0c4d564d6ae712c0"} Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.830100 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-d45mm" event={"ID":"fee2c65e-0c82-4258-8263-3de044da4402","Type":"ContainerStarted","Data":"a6074f7996d62eca3feb85117ab5e990011b33df5b80fc761b517f99e148b7e7"} Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.831390 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06da91c4-1f07-43da-9462-1bc43fb8e255","Type":"ContainerStarted","Data":"726c22e1ea8e8427b1a09e37c542133371cf2bcb74e1104fb03062dcc561e09e"} Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.835660 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerStarted","Data":"e2c10f5d25ff1a382bff2bc8aeff92ff1ca5ac2eeef2eb5b319866ec7a12ff10"} Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.836295 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.837663 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218","Type":"ContainerStarted","Data":"36380df8e1a7af246a193138757eb8fd543ae3697add37578503345fc5642a1b"} Nov 23 07:10:25 crc kubenswrapper[4906]: I1123 07:10:25.899980 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.626452966 podStartE2EDuration="5.899947841s" podCreationTimestamp="2025-11-23 07:10:20 +0000 UTC" firstStartedPulling="2025-11-23 07:10:21.554501829 +0000 UTC m=+1237.067893132" lastFinishedPulling="2025-11-23 07:10:24.827996704 +0000 UTC m=+1240.341388007" observedRunningTime="2025-11-23 07:10:25.879099655 +0000 UTC m=+1241.392490968" watchObservedRunningTime="2025-11-23 07:10:25.899947841 +0000 UTC m=+1241.413339154" Nov 23 07:10:26 crc kubenswrapper[4906]: I1123 07:10:26.602647 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:26 crc kubenswrapper[4906]: I1123 07:10:26.851673 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218","Type":"ContainerStarted","Data":"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079"} Nov 23 07:10:26 crc kubenswrapper[4906]: I1123 07:10:26.856332 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-d45mm" event={"ID":"fee2c65e-0c82-4258-8263-3de044da4402","Type":"ContainerStarted","Data":"a66239351e06dfa141b579065f17e9ef73f39c529b916a1ac8499620e7915c6d"} Nov 23 07:10:26 crc kubenswrapper[4906]: I1123 07:10:26.856451 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:26 crc kubenswrapper[4906]: I1123 07:10:26.885887 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-797bbc649-d45mm" podStartSLOduration=2.885867195 podStartE2EDuration="2.885867195s" podCreationTimestamp="2025-11-23 07:10:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:26.87957896 +0000 UTC m=+1242.392970263" watchObservedRunningTime="2025-11-23 07:10:26.885867195 +0000 UTC m=+1242.399258488" Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.106508 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.168360 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.199611 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-68644c84-m657h"] Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.199893 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-68644c84-m657h" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api-log" containerID="cri-o://c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb" gracePeriod=30 Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.200385 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-68644c84-m657h" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api" containerID="cri-o://bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094" gracePeriod=30 Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.882693 4906 generic.go:334] "Generic (PLEG): container finished" podID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerID="c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb" exitCode=143 Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.882722 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68644c84-m657h" event={"ID":"ccf9323c-773e-49e4-8788-0cc4fddcbb09","Type":"ContainerDied","Data":"c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb"} Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.889060 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218","Type":"ContainerStarted","Data":"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e"} Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.889248 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.894038 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06da91c4-1f07-43da-9462-1bc43fb8e255","Type":"ContainerStarted","Data":"ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92"} Nov 23 07:10:27 crc kubenswrapper[4906]: I1123 07:10:27.917264 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.91724441 podStartE2EDuration="3.91724441s" podCreationTimestamp="2025-11-23 07:10:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:27.913978264 +0000 UTC m=+1243.427369567" watchObservedRunningTime="2025-11-23 07:10:27.91724441 +0000 UTC m=+1243.430635713" Nov 23 07:10:28 crc kubenswrapper[4906]: I1123 07:10:28.904533 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api-log" containerID="cri-o://0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079" gracePeriod=30 Nov 23 07:10:28 crc kubenswrapper[4906]: I1123 07:10:28.906402 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06da91c4-1f07-43da-9462-1bc43fb8e255","Type":"ContainerStarted","Data":"75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d"} Nov 23 07:10:28 crc kubenswrapper[4906]: I1123 07:10:28.906823 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api" containerID="cri-o://d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e" gracePeriod=30 Nov 23 07:10:28 crc kubenswrapper[4906]: I1123 07:10:28.934104 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.028153136 podStartE2EDuration="4.934074853s" podCreationTimestamp="2025-11-23 07:10:24 +0000 UTC" firstStartedPulling="2025-11-23 07:10:24.959179971 +0000 UTC m=+1240.472571274" lastFinishedPulling="2025-11-23 07:10:25.865101688 +0000 UTC m=+1241.378492991" observedRunningTime="2025-11-23 07:10:28.928370404 +0000 UTC m=+1244.441761707" watchObservedRunningTime="2025-11-23 07:10:28.934074853 +0000 UTC m=+1244.447466196" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.423754 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.485661 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.529463 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-logs\") pod \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.529573 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-combined-ca-bundle\") pod \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.529696 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data\") pod \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.529762 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data-custom\") pod \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.529829 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmjmq\" (UniqueName: \"kubernetes.io/projected/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-kube-api-access-zmjmq\") pod \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.529910 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-etc-machine-id\") pod \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.530061 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-scripts\") pod \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\" (UID: \"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218\") " Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.532295 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" (UID: "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.533800 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-logs" (OuterVolumeSpecName: "logs") pod "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" (UID: "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.542184 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-kube-api-access-zmjmq" (OuterVolumeSpecName: "kube-api-access-zmjmq") pod "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" (UID: "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218"). InnerVolumeSpecName "kube-api-access-zmjmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.543242 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-scripts" (OuterVolumeSpecName: "scripts") pod "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" (UID: "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.545091 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" (UID: "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.591840 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data" (OuterVolumeSpecName: "config-data") pod "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" (UID: "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.592425 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" (UID: "65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.635924 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.635991 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmjmq\" (UniqueName: \"kubernetes.io/projected/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-kube-api-access-zmjmq\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.636016 4906 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.636036 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.636059 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.636076 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.636094 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.917571 4906 generic.go:334] "Generic (PLEG): container finished" podID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerID="d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e" exitCode=0 Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.917631 4906 generic.go:334] "Generic (PLEG): container finished" podID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerID="0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079" exitCode=143 Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.921313 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.930576 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218","Type":"ContainerDied","Data":"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e"} Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.930714 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218","Type":"ContainerDied","Data":"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079"} Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.930753 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218","Type":"ContainerDied","Data":"36380df8e1a7af246a193138757eb8fd543ae3697add37578503345fc5642a1b"} Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.930808 4906 scope.go:117] "RemoveContainer" containerID="d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e" Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.971429 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:29 crc kubenswrapper[4906]: I1123 07:10:29.979230 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.008346 4906 scope.go:117] "RemoveContainer" containerID="0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.013811 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:30 crc kubenswrapper[4906]: E1123 07:10:30.014730 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api-log" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.014753 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api-log" Nov 23 07:10:30 crc kubenswrapper[4906]: E1123 07:10:30.014872 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.014886 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.015288 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api-log" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.015365 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" containerName="cinder-api" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.017662 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.021114 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.022722 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.032989 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.035241 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.046621 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.046698 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.046789 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gch57\" (UniqueName: \"kubernetes.io/projected/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-kube-api-access-gch57\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.046831 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-scripts\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.046859 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.046886 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-logs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.046958 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.047003 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.047036 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.071418 4906 scope.go:117] "RemoveContainer" containerID="d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e" Nov 23 07:10:30 crc kubenswrapper[4906]: E1123 07:10:30.072318 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e\": container with ID starting with d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e not found: ID does not exist" containerID="d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.072355 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e"} err="failed to get container status \"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e\": rpc error: code = NotFound desc = could not find container \"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e\": container with ID starting with d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e not found: ID does not exist" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.072402 4906 scope.go:117] "RemoveContainer" containerID="0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079" Nov 23 07:10:30 crc kubenswrapper[4906]: E1123 07:10:30.072832 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079\": container with ID starting with 0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079 not found: ID does not exist" containerID="0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.072878 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079"} err="failed to get container status \"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079\": rpc error: code = NotFound desc = could not find container \"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079\": container with ID starting with 0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079 not found: ID does not exist" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.072896 4906 scope.go:117] "RemoveContainer" containerID="d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.073126 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e"} err="failed to get container status \"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e\": rpc error: code = NotFound desc = could not find container \"d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e\": container with ID starting with d931f0d52405426c14ea80fc864a4257c09e6d727d5e97e94b85dc600bff038e not found: ID does not exist" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.073146 4906 scope.go:117] "RemoveContainer" containerID="0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.073700 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079"} err="failed to get container status \"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079\": rpc error: code = NotFound desc = could not find container \"0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079\": container with ID starting with 0b20b37cc0356d7a4bfafe961d91cde86a929f21d2e11cfb05a0e8ae11276079 not found: ID does not exist" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.148928 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149002 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149028 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149056 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149091 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149157 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gch57\" (UniqueName: \"kubernetes.io/projected/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-kube-api-access-gch57\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149188 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-scripts\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149212 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149239 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-logs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149409 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.149763 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-logs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.153588 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-scripts\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.154757 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.155441 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.156824 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.157190 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.163696 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.168067 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gch57\" (UniqueName: \"kubernetes.io/projected/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-kube-api-access-gch57\") pod \"cinder-api-0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.378861 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.827467 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-68644c84-m657h" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:43054->10.217.0.154:9311: read: connection reset by peer" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.827477 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-68644c84-m657h" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:43068->10.217.0.154:9311: read: connection reset by peer" Nov 23 07:10:30 crc kubenswrapper[4906]: I1123 07:10:30.897495 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:10:30 crc kubenswrapper[4906]: W1123 07:10:30.927302 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ee97571_4ae5_4ed0_a32a_67abd5896ae0.slice/crio-b8acffca0291884aad6ad380d3e1bceec373abb266c89d76eeec84689419d1b3 WatchSource:0}: Error finding container b8acffca0291884aad6ad380d3e1bceec373abb266c89d76eeec84689419d1b3: Status 404 returned error can't find the container with id b8acffca0291884aad6ad380d3e1bceec373abb266c89d76eeec84689419d1b3 Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.324013 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.373878 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218" path="/var/lib/kubelet/pods/65ca8ef2-8cdd-4b3f-9705-03fdbbdcd218/volumes" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.385536 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccf9323c-773e-49e4-8788-0cc4fddcbb09-logs\") pod \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.385736 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-combined-ca-bundle\") pod \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.385865 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data-custom\") pod \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.385916 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data\") pod \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.385977 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s655\" (UniqueName: \"kubernetes.io/projected/ccf9323c-773e-49e4-8788-0cc4fddcbb09-kube-api-access-5s655\") pod \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\" (UID: \"ccf9323c-773e-49e4-8788-0cc4fddcbb09\") " Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.389009 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccf9323c-773e-49e4-8788-0cc4fddcbb09-logs" (OuterVolumeSpecName: "logs") pod "ccf9323c-773e-49e4-8788-0cc4fddcbb09" (UID: "ccf9323c-773e-49e4-8788-0cc4fddcbb09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.416110 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccf9323c-773e-49e4-8788-0cc4fddcbb09-kube-api-access-5s655" (OuterVolumeSpecName: "kube-api-access-5s655") pod "ccf9323c-773e-49e4-8788-0cc4fddcbb09" (UID: "ccf9323c-773e-49e4-8788-0cc4fddcbb09"). InnerVolumeSpecName "kube-api-access-5s655". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.421129 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ccf9323c-773e-49e4-8788-0cc4fddcbb09" (UID: "ccf9323c-773e-49e4-8788-0cc4fddcbb09"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.421181 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccf9323c-773e-49e4-8788-0cc4fddcbb09" (UID: "ccf9323c-773e-49e4-8788-0cc4fddcbb09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.454126 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data" (OuterVolumeSpecName: "config-data") pod "ccf9323c-773e-49e4-8788-0cc4fddcbb09" (UID: "ccf9323c-773e-49e4-8788-0cc4fddcbb09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.488564 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccf9323c-773e-49e4-8788-0cc4fddcbb09-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.488598 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.488612 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.488622 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccf9323c-773e-49e4-8788-0cc4fddcbb09-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.488631 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s655\" (UniqueName: \"kubernetes.io/projected/ccf9323c-773e-49e4-8788-0cc4fddcbb09-kube-api-access-5s655\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.645395 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.962195 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ee97571-4ae5-4ed0-a32a-67abd5896ae0","Type":"ContainerStarted","Data":"df25a67d5a9f5925815eee2060a4fdcc509bfcf08b8d1d1a7f6827a40492c808"} Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.962252 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ee97571-4ae5-4ed0-a32a-67abd5896ae0","Type":"ContainerStarted","Data":"b8acffca0291884aad6ad380d3e1bceec373abb266c89d76eeec84689419d1b3"} Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.968349 4906 generic.go:334] "Generic (PLEG): container finished" podID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerID="bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094" exitCode=0 Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.968382 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68644c84-m657h" event={"ID":"ccf9323c-773e-49e4-8788-0cc4fddcbb09","Type":"ContainerDied","Data":"bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094"} Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.968401 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68644c84-m657h" event={"ID":"ccf9323c-773e-49e4-8788-0cc4fddcbb09","Type":"ContainerDied","Data":"2838b650edab37ea36c04e8bea9f5159a43698a221607fc325214688b1e6fc31"} Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.968418 4906 scope.go:117] "RemoveContainer" containerID="bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094" Nov 23 07:10:31 crc kubenswrapper[4906]: I1123 07:10:31.968591 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68644c84-m657h" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.016840 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-68644c84-m657h"] Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.022782 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-68644c84-m657h"] Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.035338 4906 scope.go:117] "RemoveContainer" containerID="c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.075936 4906 scope.go:117] "RemoveContainer" containerID="bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094" Nov 23 07:10:32 crc kubenswrapper[4906]: E1123 07:10:32.076419 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094\": container with ID starting with bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094 not found: ID does not exist" containerID="bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.076459 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094"} err="failed to get container status \"bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094\": rpc error: code = NotFound desc = could not find container \"bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094\": container with ID starting with bcdb66cc5352dc6c77edc6f83e56f861167a64667b294f37ac2917e1fece3094 not found: ID does not exist" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.076482 4906 scope.go:117] "RemoveContainer" containerID="c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb" Nov 23 07:10:32 crc kubenswrapper[4906]: E1123 07:10:32.079860 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb\": container with ID starting with c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb not found: ID does not exist" containerID="c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.079917 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb"} err="failed to get container status \"c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb\": rpc error: code = NotFound desc = could not find container \"c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb\": container with ID starting with c4055e8587084ce2828251c81dc75eb869cd46c8537d5f1c8f6aee1bebd16efb not found: ID does not exist" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.263222 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.895382 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.980081 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ee97571-4ae5-4ed0-a32a-67abd5896ae0","Type":"ContainerStarted","Data":"e82535f708e61c7ee739e897ea48ec09a393ccd402bebffaf681b12cea2dcb54"} Nov 23 07:10:32 crc kubenswrapper[4906]: I1123 07:10:32.980502 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 23 07:10:33 crc kubenswrapper[4906]: I1123 07:10:33.004240 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.004218122 podStartE2EDuration="4.004218122s" podCreationTimestamp="2025-11-23 07:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:32.997673471 +0000 UTC m=+1248.511064774" watchObservedRunningTime="2025-11-23 07:10:33.004218122 +0000 UTC m=+1248.517609425" Nov 23 07:10:33 crc kubenswrapper[4906]: I1123 07:10:33.371534 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" path="/var/lib/kubelet/pods/ccf9323c-773e-49e4-8788-0cc4fddcbb09/volumes" Nov 23 07:10:34 crc kubenswrapper[4906]: I1123 07:10:34.501975 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:10:34 crc kubenswrapper[4906]: I1123 07:10:34.592267 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-z9rz8"] Nov 23 07:10:34 crc kubenswrapper[4906]: I1123 07:10:34.592805 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" podUID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerName="dnsmasq-dns" containerID="cri-o://e9a97e802ec3d2f28e7ab2b9a9e4811701dbf5d52cbc5746484f45eee01458d6" gracePeriod=10 Nov 23 07:10:34 crc kubenswrapper[4906]: I1123 07:10:34.764137 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 23 07:10:34 crc kubenswrapper[4906]: I1123 07:10:34.883078 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.022911 4906 generic.go:334] "Generic (PLEG): container finished" podID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerID="e9a97e802ec3d2f28e7ab2b9a9e4811701dbf5d52cbc5746484f45eee01458d6" exitCode=0 Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.023289 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="cinder-scheduler" containerID="cri-o://ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92" gracePeriod=30 Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.023444 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="probe" containerID="cri-o://75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d" gracePeriod=30 Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.023695 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" event={"ID":"495c2958-59c5-4eb7-a9ee-299e675cf8ff","Type":"ContainerDied","Data":"e9a97e802ec3d2f28e7ab2b9a9e4811701dbf5d52cbc5746484f45eee01458d6"} Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.246951 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.283982 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-svc\") pod \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.284047 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-swift-storage-0\") pod \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.284108 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-config\") pod \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.284156 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-sb\") pod \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.284183 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zpd5\" (UniqueName: \"kubernetes.io/projected/495c2958-59c5-4eb7-a9ee-299e675cf8ff-kube-api-access-2zpd5\") pod \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.284231 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-nb\") pod \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\" (UID: \"495c2958-59c5-4eb7-a9ee-299e675cf8ff\") " Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.298175 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495c2958-59c5-4eb7-a9ee-299e675cf8ff-kube-api-access-2zpd5" (OuterVolumeSpecName: "kube-api-access-2zpd5") pod "495c2958-59c5-4eb7-a9ee-299e675cf8ff" (UID: "495c2958-59c5-4eb7-a9ee-299e675cf8ff"). InnerVolumeSpecName "kube-api-access-2zpd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.390966 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zpd5\" (UniqueName: \"kubernetes.io/projected/495c2958-59c5-4eb7-a9ee-299e675cf8ff-kube-api-access-2zpd5\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.440186 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "495c2958-59c5-4eb7-a9ee-299e675cf8ff" (UID: "495c2958-59c5-4eb7-a9ee-299e675cf8ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.476353 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "495c2958-59c5-4eb7-a9ee-299e675cf8ff" (UID: "495c2958-59c5-4eb7-a9ee-299e675cf8ff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.493362 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.493395 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.494262 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-config" (OuterVolumeSpecName: "config") pod "495c2958-59c5-4eb7-a9ee-299e675cf8ff" (UID: "495c2958-59c5-4eb7-a9ee-299e675cf8ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.504391 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "495c2958-59c5-4eb7-a9ee-299e675cf8ff" (UID: "495c2958-59c5-4eb7-a9ee-299e675cf8ff"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.525595 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "495c2958-59c5-4eb7-a9ee-299e675cf8ff" (UID: "495c2958-59c5-4eb7-a9ee-299e675cf8ff"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.596133 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.596479 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.596557 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/495c2958-59c5-4eb7-a9ee-299e675cf8ff-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.727228 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.817896 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56d695cc56-fkbp4"] Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.818598 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56d695cc56-fkbp4" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-api" containerID="cri-o://c0578e8d6b5b396123727d6c012bf4d48a3e9c86ec3bea27282318f3e04bfffa" gracePeriod=30 Nov 23 07:10:35 crc kubenswrapper[4906]: I1123 07:10:35.819190 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56d695cc56-fkbp4" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-httpd" containerID="cri-o://ad06d562f80a95bcf1c5e5404e0c38aa2d345c32e7bce7a62c5cf872ff3fe9f4" gracePeriod=30 Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.051807 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.051791 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-z9rz8" event={"ID":"495c2958-59c5-4eb7-a9ee-299e675cf8ff","Type":"ContainerDied","Data":"cd04ecab952d4e13899cfb8e0f1d5e8081956cb0882d142a26d46f88012d1a43"} Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.052072 4906 scope.go:117] "RemoveContainer" containerID="e9a97e802ec3d2f28e7ab2b9a9e4811701dbf5d52cbc5746484f45eee01458d6" Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.069410 4906 generic.go:334] "Generic (PLEG): container finished" podID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerID="75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d" exitCode=0 Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.069465 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06da91c4-1f07-43da-9462-1bc43fb8e255","Type":"ContainerDied","Data":"75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d"} Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.105197 4906 scope.go:117] "RemoveContainer" containerID="b084e4cd79cd99c0619e4f93a8893304d5cf1f0f0eefbcdca66d25decbb61cda" Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.114978 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-z9rz8"] Nov 23 07:10:36 crc kubenswrapper[4906]: I1123 07:10:36.131898 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-z9rz8"] Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.051765 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 23 07:10:37 crc kubenswrapper[4906]: E1123 07:10:37.052704 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerName="init" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.052720 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerName="init" Nov 23 07:10:37 crc kubenswrapper[4906]: E1123 07:10:37.052739 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerName="dnsmasq-dns" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.052745 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerName="dnsmasq-dns" Nov 23 07:10:37 crc kubenswrapper[4906]: E1123 07:10:37.052768 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.052774 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api" Nov 23 07:10:37 crc kubenswrapper[4906]: E1123 07:10:37.052785 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api-log" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.052792 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api-log" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.052970 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api-log" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.052996 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" containerName="dnsmasq-dns" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.053008 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf9323c-773e-49e4-8788-0cc4fddcbb09" containerName="barbican-api" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.053758 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.066242 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bslpb" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.066532 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.066538 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.080347 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.123482 4906 generic.go:334] "Generic (PLEG): container finished" podID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerID="ad06d562f80a95bcf1c5e5404e0c38aa2d345c32e7bce7a62c5cf872ff3fe9f4" exitCode=0 Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.123549 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d695cc56-fkbp4" event={"ID":"00e11c87-d705-4b7a-8771-13d3117e67e7","Type":"ContainerDied","Data":"ad06d562f80a95bcf1c5e5404e0c38aa2d345c32e7bce7a62c5cf872ff3fe9f4"} Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.133603 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config-secret\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.133644 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.133695 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfzfv\" (UniqueName: \"kubernetes.io/projected/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-kube-api-access-rfzfv\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.133813 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.236762 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfzfv\" (UniqueName: \"kubernetes.io/projected/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-kube-api-access-rfzfv\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.236828 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.236971 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config-secret\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.236992 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.238167 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.248469 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.257146 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfzfv\" (UniqueName: \"kubernetes.io/projected/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-kube-api-access-rfzfv\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.269894 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config-secret\") pod \"openstackclient\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.390530 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="495c2958-59c5-4eb7-a9ee-299e675cf8ff" path="/var/lib/kubelet/pods/495c2958-59c5-4eb7-a9ee-299e675cf8ff/volumes" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.396282 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:10:37 crc kubenswrapper[4906]: I1123 07:10:37.946948 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 07:10:38 crc kubenswrapper[4906]: I1123 07:10:38.141610 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05","Type":"ContainerStarted","Data":"6fc541f1fefd3436a7962b1f8796f14ec3be58de18f362b3023a9831bd73bce1"} Nov 23 07:10:38 crc kubenswrapper[4906]: I1123 07:10:38.969528 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.081321 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jvpn\" (UniqueName: \"kubernetes.io/projected/06da91c4-1f07-43da-9462-1bc43fb8e255-kube-api-access-8jvpn\") pod \"06da91c4-1f07-43da-9462-1bc43fb8e255\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.081379 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-scripts\") pod \"06da91c4-1f07-43da-9462-1bc43fb8e255\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.081437 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06da91c4-1f07-43da-9462-1bc43fb8e255-etc-machine-id\") pod \"06da91c4-1f07-43da-9462-1bc43fb8e255\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.081494 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data-custom\") pod \"06da91c4-1f07-43da-9462-1bc43fb8e255\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.081524 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data\") pod \"06da91c4-1f07-43da-9462-1bc43fb8e255\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.081570 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-combined-ca-bundle\") pod \"06da91c4-1f07-43da-9462-1bc43fb8e255\" (UID: \"06da91c4-1f07-43da-9462-1bc43fb8e255\") " Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.081707 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06da91c4-1f07-43da-9462-1bc43fb8e255-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "06da91c4-1f07-43da-9462-1bc43fb8e255" (UID: "06da91c4-1f07-43da-9462-1bc43fb8e255"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.082001 4906 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06da91c4-1f07-43da-9462-1bc43fb8e255-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.089234 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-scripts" (OuterVolumeSpecName: "scripts") pod "06da91c4-1f07-43da-9462-1bc43fb8e255" (UID: "06da91c4-1f07-43da-9462-1bc43fb8e255"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.089317 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "06da91c4-1f07-43da-9462-1bc43fb8e255" (UID: "06da91c4-1f07-43da-9462-1bc43fb8e255"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.092985 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06da91c4-1f07-43da-9462-1bc43fb8e255-kube-api-access-8jvpn" (OuterVolumeSpecName: "kube-api-access-8jvpn") pod "06da91c4-1f07-43da-9462-1bc43fb8e255" (UID: "06da91c4-1f07-43da-9462-1bc43fb8e255"). InnerVolumeSpecName "kube-api-access-8jvpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.153729 4906 generic.go:334] "Generic (PLEG): container finished" podID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerID="ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92" exitCode=0 Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.153792 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06da91c4-1f07-43da-9462-1bc43fb8e255","Type":"ContainerDied","Data":"ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92"} Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.153878 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06da91c4-1f07-43da-9462-1bc43fb8e255","Type":"ContainerDied","Data":"726c22e1ea8e8427b1a09e37c542133371cf2bcb74e1104fb03062dcc561e09e"} Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.153904 4906 scope.go:117] "RemoveContainer" containerID="75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.153910 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.186793 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jvpn\" (UniqueName: \"kubernetes.io/projected/06da91c4-1f07-43da-9462-1bc43fb8e255-kube-api-access-8jvpn\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.186828 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.186841 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.192010 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06da91c4-1f07-43da-9462-1bc43fb8e255" (UID: "06da91c4-1f07-43da-9462-1bc43fb8e255"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.197800 4906 scope.go:117] "RemoveContainer" containerID="ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.207090 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data" (OuterVolumeSpecName: "config-data") pod "06da91c4-1f07-43da-9462-1bc43fb8e255" (UID: "06da91c4-1f07-43da-9462-1bc43fb8e255"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.225166 4906 scope.go:117] "RemoveContainer" containerID="75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d" Nov 23 07:10:39 crc kubenswrapper[4906]: E1123 07:10:39.225920 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d\": container with ID starting with 75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d not found: ID does not exist" containerID="75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.225997 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d"} err="failed to get container status \"75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d\": rpc error: code = NotFound desc = could not find container \"75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d\": container with ID starting with 75f63d7a9082046fb1757f7ace0067666d47d4c901c6d5d158c070bf92dbf57d not found: ID does not exist" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.226055 4906 scope.go:117] "RemoveContainer" containerID="ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92" Nov 23 07:10:39 crc kubenswrapper[4906]: E1123 07:10:39.226502 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92\": container with ID starting with ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92 not found: ID does not exist" containerID="ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.226557 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92"} err="failed to get container status \"ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92\": rpc error: code = NotFound desc = could not find container \"ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92\": container with ID starting with ad3dc0b8cdde5096863352ad0004884650172acdbbb4869b8d69208dc9a4ac92 not found: ID does not exist" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.289235 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.289280 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06da91c4-1f07-43da-9462-1bc43fb8e255-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.514398 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.537757 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.550735 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:39 crc kubenswrapper[4906]: E1123 07:10:39.551242 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="probe" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.551263 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="probe" Nov 23 07:10:39 crc kubenswrapper[4906]: E1123 07:10:39.551274 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="cinder-scheduler" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.551281 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="cinder-scheduler" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.551517 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="probe" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.551539 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" containerName="cinder-scheduler" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.552760 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.555730 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.560542 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.596822 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.597434 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.597603 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.597820 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxt8r\" (UniqueName: \"kubernetes.io/projected/a7788825-483c-46f2-a975-239254435019-kube-api-access-jxt8r\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.598004 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7788825-483c-46f2-a975-239254435019-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.598117 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-scripts\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.701232 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.701573 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.701797 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.701890 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxt8r\" (UniqueName: \"kubernetes.io/projected/a7788825-483c-46f2-a975-239254435019-kube-api-access-jxt8r\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.701945 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7788825-483c-46f2-a975-239254435019-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.702013 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-scripts\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.702169 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7788825-483c-46f2-a975-239254435019-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.706467 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-scripts\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.710748 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.711652 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.713483 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.731450 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxt8r\" (UniqueName: \"kubernetes.io/projected/a7788825-483c-46f2-a975-239254435019-kube-api-access-jxt8r\") pod \"cinder-scheduler-0\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " pod="openstack/cinder-scheduler-0" Nov 23 07:10:39 crc kubenswrapper[4906]: I1123 07:10:39.868418 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:10:40 crc kubenswrapper[4906]: I1123 07:10:40.502628 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:10:40 crc kubenswrapper[4906]: W1123 07:10:40.509754 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7788825_483c_46f2_a975_239254435019.slice/crio-55ac9528202f43bf2d8a88071468bd7753abea70b0cbb303b4d4b35b5acac902 WatchSource:0}: Error finding container 55ac9528202f43bf2d8a88071468bd7753abea70b0cbb303b4d4b35b5acac902: Status 404 returned error can't find the container with id 55ac9528202f43bf2d8a88071468bd7753abea70b0cbb303b4d4b35b5acac902 Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.015166 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-77c4d866c-r9x72"] Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.022763 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.025838 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.026097 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.032946 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-77c4d866c-r9x72"] Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.046425 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.146435 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-public-tls-certs\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.146525 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-run-httpd\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.146569 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-combined-ca-bundle\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.146590 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-config-data\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.146811 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-etc-swift\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.147033 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-log-httpd\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.147197 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-internal-tls-certs\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.147337 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfrqx\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-kube-api-access-jfrqx\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.231779 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7788825-483c-46f2-a975-239254435019","Type":"ContainerStarted","Data":"55ac9528202f43bf2d8a88071468bd7753abea70b0cbb303b4d4b35b5acac902"} Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.249666 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-log-httpd\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.249783 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-internal-tls-certs\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.249858 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfrqx\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-kube-api-access-jfrqx\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.249972 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-public-tls-certs\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.250067 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-run-httpd\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.250102 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-combined-ca-bundle\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.250121 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-config-data\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.250191 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-etc-swift\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.259175 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-run-httpd\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.264392 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-config-data\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.264516 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-etc-swift\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.269268 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-public-tls-certs\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.278390 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-internal-tls-certs\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.279706 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfrqx\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-kube-api-access-jfrqx\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.281136 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-log-httpd\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.283337 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-combined-ca-bundle\") pod \"swift-proxy-77c4d866c-r9x72\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.354956 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:41 crc kubenswrapper[4906]: I1123 07:10:41.373063 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06da91c4-1f07-43da-9462-1bc43fb8e255" path="/var/lib/kubelet/pods/06da91c4-1f07-43da-9462-1bc43fb8e255/volumes" Nov 23 07:10:42 crc kubenswrapper[4906]: I1123 07:10:42.208104 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-77c4d866c-r9x72"] Nov 23 07:10:42 crc kubenswrapper[4906]: I1123 07:10:42.254608 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77c4d866c-r9x72" event={"ID":"b4121204-6477-44c6-b669-fc7497fa908d","Type":"ContainerStarted","Data":"e6db7921dc7dbf6ed9d3d94c5afccf6bd7f05007a7f4deb53dcdac820859140f"} Nov 23 07:10:42 crc kubenswrapper[4906]: I1123 07:10:42.266000 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7788825-483c-46f2-a975-239254435019","Type":"ContainerStarted","Data":"209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d"} Nov 23 07:10:42 crc kubenswrapper[4906]: I1123 07:10:42.761160 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 23 07:10:43 crc kubenswrapper[4906]: I1123 07:10:43.285719 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77c4d866c-r9x72" event={"ID":"b4121204-6477-44c6-b669-fc7497fa908d","Type":"ContainerStarted","Data":"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113"} Nov 23 07:10:43 crc kubenswrapper[4906]: I1123 07:10:43.291580 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77c4d866c-r9x72" event={"ID":"b4121204-6477-44c6-b669-fc7497fa908d","Type":"ContainerStarted","Data":"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7"} Nov 23 07:10:43 crc kubenswrapper[4906]: I1123 07:10:43.291661 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:43 crc kubenswrapper[4906]: I1123 07:10:43.291675 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:43 crc kubenswrapper[4906]: I1123 07:10:43.296426 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7788825-483c-46f2-a975-239254435019","Type":"ContainerStarted","Data":"9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c"} Nov 23 07:10:43 crc kubenswrapper[4906]: I1123 07:10:43.327504 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-77c4d866c-r9x72" podStartSLOduration=3.32747754 podStartE2EDuration="3.32747754s" podCreationTimestamp="2025-11-23 07:10:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:43.315388754 +0000 UTC m=+1258.828780057" watchObservedRunningTime="2025-11-23 07:10:43.32747754 +0000 UTC m=+1258.840868843" Nov 23 07:10:43 crc kubenswrapper[4906]: I1123 07:10:43.344318 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.34427463 podStartE2EDuration="4.34427463s" podCreationTimestamp="2025-11-23 07:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:10:43.341296122 +0000 UTC m=+1258.854687425" watchObservedRunningTime="2025-11-23 07:10:43.34427463 +0000 UTC m=+1258.857665933" Nov 23 07:10:44 crc kubenswrapper[4906]: I1123 07:10:44.526535 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:44 crc kubenswrapper[4906]: I1123 07:10:44.527309 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-central-agent" containerID="cri-o://66dff64417cdf5fe7d1ce3e986378bd84df2c47d8565a82aeda164822b6780db" gracePeriod=30 Nov 23 07:10:44 crc kubenswrapper[4906]: I1123 07:10:44.527380 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="proxy-httpd" containerID="cri-o://e2c10f5d25ff1a382bff2bc8aeff92ff1ca5ac2eeef2eb5b319866ec7a12ff10" gracePeriod=30 Nov 23 07:10:44 crc kubenswrapper[4906]: I1123 07:10:44.527441 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="sg-core" containerID="cri-o://f25969cc9818ffc774942bd7e1fc0db8a15e76f95b2f056249d5b1eec853d7e2" gracePeriod=30 Nov 23 07:10:44 crc kubenswrapper[4906]: I1123 07:10:44.527488 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-notification-agent" containerID="cri-o://122f5ddae4d56f6dd982766ee340c268fe423822feea55f993b8a2cb95d2b4a4" gracePeriod=30 Nov 23 07:10:44 crc kubenswrapper[4906]: I1123 07:10:44.638080 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.157:3000/\": read tcp 10.217.0.2:56052->10.217.0.157:3000: read: connection reset by peer" Nov 23 07:10:44 crc kubenswrapper[4906]: I1123 07:10:44.869132 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 23 07:10:45 crc kubenswrapper[4906]: I1123 07:10:45.349150 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerDied","Data":"e2c10f5d25ff1a382bff2bc8aeff92ff1ca5ac2eeef2eb5b319866ec7a12ff10"} Nov 23 07:10:45 crc kubenswrapper[4906]: I1123 07:10:45.348956 4906 generic.go:334] "Generic (PLEG): container finished" podID="4533768d-d850-457e-87cf-7adf0d594915" containerID="e2c10f5d25ff1a382bff2bc8aeff92ff1ca5ac2eeef2eb5b319866ec7a12ff10" exitCode=0 Nov 23 07:10:45 crc kubenswrapper[4906]: I1123 07:10:45.349292 4906 generic.go:334] "Generic (PLEG): container finished" podID="4533768d-d850-457e-87cf-7adf0d594915" containerID="f25969cc9818ffc774942bd7e1fc0db8a15e76f95b2f056249d5b1eec853d7e2" exitCode=2 Nov 23 07:10:45 crc kubenswrapper[4906]: I1123 07:10:45.349321 4906 generic.go:334] "Generic (PLEG): container finished" podID="4533768d-d850-457e-87cf-7adf0d594915" containerID="66dff64417cdf5fe7d1ce3e986378bd84df2c47d8565a82aeda164822b6780db" exitCode=0 Nov 23 07:10:45 crc kubenswrapper[4906]: I1123 07:10:45.350169 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerDied","Data":"f25969cc9818ffc774942bd7e1fc0db8a15e76f95b2f056249d5b1eec853d7e2"} Nov 23 07:10:45 crc kubenswrapper[4906]: I1123 07:10:45.350228 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerDied","Data":"66dff64417cdf5fe7d1ce3e986378bd84df2c47d8565a82aeda164822b6780db"} Nov 23 07:10:46 crc kubenswrapper[4906]: I1123 07:10:46.369132 4906 generic.go:334] "Generic (PLEG): container finished" podID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerID="c0578e8d6b5b396123727d6c012bf4d48a3e9c86ec3bea27282318f3e04bfffa" exitCode=0 Nov 23 07:10:46 crc kubenswrapper[4906]: I1123 07:10:46.369283 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d695cc56-fkbp4" event={"ID":"00e11c87-d705-4b7a-8771-13d3117e67e7","Type":"ContainerDied","Data":"c0578e8d6b5b396123727d6c012bf4d48a3e9c86ec3bea27282318f3e04bfffa"} Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.072045 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.423129 4906 generic.go:334] "Generic (PLEG): container finished" podID="4533768d-d850-457e-87cf-7adf0d594915" containerID="122f5ddae4d56f6dd982766ee340c268fe423822feea55f993b8a2cb95d2b4a4" exitCode=0 Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.423181 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerDied","Data":"122f5ddae4d56f6dd982766ee340c268fe423822feea55f993b8a2cb95d2b4a4"} Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.946042 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.946583 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.946649 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.947886 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cde0743cc2b7d9547c3784b89c951ad5af5f1a5ec9c54292ba0b98af998c4c47"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:10:50 crc kubenswrapper[4906]: I1123 07:10:50.947972 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://cde0743cc2b7d9547c3784b89c951ad5af5f1a5ec9c54292ba0b98af998c4c47" gracePeriod=600 Nov 23 07:10:51 crc kubenswrapper[4906]: I1123 07:10:51.100384 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.157:3000/\": dial tcp 10.217.0.157:3000: connect: connection refused" Nov 23 07:10:51 crc kubenswrapper[4906]: I1123 07:10:51.371526 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:51 crc kubenswrapper[4906]: I1123 07:10:51.373714 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:10:51 crc kubenswrapper[4906]: I1123 07:10:51.443294 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="cde0743cc2b7d9547c3784b89c951ad5af5f1a5ec9c54292ba0b98af998c4c47" exitCode=0 Nov 23 07:10:51 crc kubenswrapper[4906]: I1123 07:10:51.443466 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"cde0743cc2b7d9547c3784b89c951ad5af5f1a5ec9c54292ba0b98af998c4c47"} Nov 23 07:10:51 crc kubenswrapper[4906]: I1123 07:10:51.443570 4906 scope.go:117] "RemoveContainer" containerID="c632a9b1605581f702cb4b0b6cbc68235936754af5501ce83c90241ccf464cb9" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.638955 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.655565 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740118 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-combined-ca-bundle\") pod \"4533768d-d850-457e-87cf-7adf0d594915\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740237 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-sg-core-conf-yaml\") pod \"4533768d-d850-457e-87cf-7adf0d594915\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740287 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-run-httpd\") pod \"4533768d-d850-457e-87cf-7adf0d594915\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740410 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-log-httpd\") pod \"4533768d-d850-457e-87cf-7adf0d594915\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740441 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-config-data\") pod \"4533768d-d850-457e-87cf-7adf0d594915\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740490 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp6ch\" (UniqueName: \"kubernetes.io/projected/4533768d-d850-457e-87cf-7adf0d594915-kube-api-access-zp6ch\") pod \"4533768d-d850-457e-87cf-7adf0d594915\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740606 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-scripts\") pod \"4533768d-d850-457e-87cf-7adf0d594915\" (UID: \"4533768d-d850-457e-87cf-7adf0d594915\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.740951 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4533768d-d850-457e-87cf-7adf0d594915" (UID: "4533768d-d850-457e-87cf-7adf0d594915"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.741148 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4533768d-d850-457e-87cf-7adf0d594915" (UID: "4533768d-d850-457e-87cf-7adf0d594915"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.741717 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.741738 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4533768d-d850-457e-87cf-7adf0d594915-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.750164 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-scripts" (OuterVolumeSpecName: "scripts") pod "4533768d-d850-457e-87cf-7adf0d594915" (UID: "4533768d-d850-457e-87cf-7adf0d594915"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.757870 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4533768d-d850-457e-87cf-7adf0d594915-kube-api-access-zp6ch" (OuterVolumeSpecName: "kube-api-access-zp6ch") pod "4533768d-d850-457e-87cf-7adf0d594915" (UID: "4533768d-d850-457e-87cf-7adf0d594915"). InnerVolumeSpecName "kube-api-access-zp6ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.776179 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4533768d-d850-457e-87cf-7adf0d594915" (UID: "4533768d-d850-457e-87cf-7adf0d594915"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.843051 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-httpd-config\") pod \"00e11c87-d705-4b7a-8771-13d3117e67e7\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.844246 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbqwv\" (UniqueName: \"kubernetes.io/projected/00e11c87-d705-4b7a-8771-13d3117e67e7-kube-api-access-wbqwv\") pod \"00e11c87-d705-4b7a-8771-13d3117e67e7\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.844340 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-combined-ca-bundle\") pod \"00e11c87-d705-4b7a-8771-13d3117e67e7\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.844422 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-ovndb-tls-certs\") pod \"00e11c87-d705-4b7a-8771-13d3117e67e7\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.844455 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-config\") pod \"00e11c87-d705-4b7a-8771-13d3117e67e7\" (UID: \"00e11c87-d705-4b7a-8771-13d3117e67e7\") " Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.845865 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp6ch\" (UniqueName: \"kubernetes.io/projected/4533768d-d850-457e-87cf-7adf0d594915-kube-api-access-zp6ch\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.845888 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.845900 4906 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.850045 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "00e11c87-d705-4b7a-8771-13d3117e67e7" (UID: "00e11c87-d705-4b7a-8771-13d3117e67e7"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.852215 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e11c87-d705-4b7a-8771-13d3117e67e7-kube-api-access-wbqwv" (OuterVolumeSpecName: "kube-api-access-wbqwv") pod "00e11c87-d705-4b7a-8771-13d3117e67e7" (UID: "00e11c87-d705-4b7a-8771-13d3117e67e7"). InnerVolumeSpecName "kube-api-access-wbqwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.853924 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4533768d-d850-457e-87cf-7adf0d594915" (UID: "4533768d-d850-457e-87cf-7adf0d594915"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.900838 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-config-data" (OuterVolumeSpecName: "config-data") pod "4533768d-d850-457e-87cf-7adf0d594915" (UID: "4533768d-d850-457e-87cf-7adf0d594915"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.915496 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00e11c87-d705-4b7a-8771-13d3117e67e7" (UID: "00e11c87-d705-4b7a-8771-13d3117e67e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.922101 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-config" (OuterVolumeSpecName: "config") pod "00e11c87-d705-4b7a-8771-13d3117e67e7" (UID: "00e11c87-d705-4b7a-8771-13d3117e67e7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.947929 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.947979 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbqwv\" (UniqueName: \"kubernetes.io/projected/00e11c87-d705-4b7a-8771-13d3117e67e7-kube-api-access-wbqwv\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.948006 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.948016 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.948026 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.948045 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4533768d-d850-457e-87cf-7adf0d594915-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:52 crc kubenswrapper[4906]: I1123 07:10:52.973068 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "00e11c87-d705-4b7a-8771-13d3117e67e7" (UID: "00e11c87-d705-4b7a-8771-13d3117e67e7"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.049174 4906 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00e11c87-d705-4b7a-8771-13d3117e67e7-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.468796 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4533768d-d850-457e-87cf-7adf0d594915","Type":"ContainerDied","Data":"557b365223e12f4583cbd2b9fcf0690a8da259f8e4e5d5b1f26befa836cbc34a"} Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.468860 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.469199 4906 scope.go:117] "RemoveContainer" containerID="e2c10f5d25ff1a382bff2bc8aeff92ff1ca5ac2eeef2eb5b319866ec7a12ff10" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.476843 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"efb82b69034455635a0b7ce63f8c53d1bea8631681ec6e17e626273b28cdbdd5"} Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.480889 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d695cc56-fkbp4" event={"ID":"00e11c87-d705-4b7a-8771-13d3117e67e7","Type":"ContainerDied","Data":"82959acd674edb822ef56fc5944fb4da0381b1425ca7deece625633c7690ccb8"} Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.481013 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d695cc56-fkbp4" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.485059 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05","Type":"ContainerStarted","Data":"6bc391c38da2e27698e0bb15d809a3a29934990091bf87fa84600dc4d231301b"} Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.489993 4906 scope.go:117] "RemoveContainer" containerID="f25969cc9818ffc774942bd7e1fc0db8a15e76f95b2f056249d5b1eec853d7e2" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.500265 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.510551 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.514867 4906 scope.go:117] "RemoveContainer" containerID="122f5ddae4d56f6dd982766ee340c268fe423822feea55f993b8a2cb95d2b4a4" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.536322 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:53 crc kubenswrapper[4906]: E1123 07:10:53.537140 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-notification-agent" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537162 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-notification-agent" Nov 23 07:10:53 crc kubenswrapper[4906]: E1123 07:10:53.537178 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="proxy-httpd" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537184 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="proxy-httpd" Nov 23 07:10:53 crc kubenswrapper[4906]: E1123 07:10:53.537213 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-central-agent" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537220 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-central-agent" Nov 23 07:10:53 crc kubenswrapper[4906]: E1123 07:10:53.537237 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-api" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537242 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-api" Nov 23 07:10:53 crc kubenswrapper[4906]: E1123 07:10:53.537256 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="sg-core" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537262 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="sg-core" Nov 23 07:10:53 crc kubenswrapper[4906]: E1123 07:10:53.537270 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-httpd" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537277 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-httpd" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537478 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-api" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537486 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-notification-agent" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537499 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" containerName="neutron-httpd" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537514 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="sg-core" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537525 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="ceilometer-central-agent" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.537538 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4533768d-d850-457e-87cf-7adf0d594915" containerName="proxy-httpd" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.539440 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.541104 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.697926963 podStartE2EDuration="16.541082014s" podCreationTimestamp="2025-11-23 07:10:37 +0000 UTC" firstStartedPulling="2025-11-23 07:10:37.957120932 +0000 UTC m=+1253.470512235" lastFinishedPulling="2025-11-23 07:10:52.800275983 +0000 UTC m=+1268.313667286" observedRunningTime="2025-11-23 07:10:53.522647811 +0000 UTC m=+1269.036039114" watchObservedRunningTime="2025-11-23 07:10:53.541082014 +0000 UTC m=+1269.054473317" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.543119 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.543770 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.558415 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.567588 4906 scope.go:117] "RemoveContainer" containerID="66dff64417cdf5fe7d1ce3e986378bd84df2c47d8565a82aeda164822b6780db" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.576566 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56d695cc56-fkbp4"] Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.586350 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-56d695cc56-fkbp4"] Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.604198 4906 scope.go:117] "RemoveContainer" containerID="ad06d562f80a95bcf1c5e5404e0c38aa2d345c32e7bce7a62c5cf872ff3fe9f4" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.629196 4906 scope.go:117] "RemoveContainer" containerID="c0578e8d6b5b396123727d6c012bf4d48a3e9c86ec3bea27282318f3e04bfffa" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.663462 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.663606 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r68z9\" (UniqueName: \"kubernetes.io/projected/4484f510-ece4-42be-b7cd-ede2e47777b1-kube-api-access-r68z9\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.664088 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-config-data\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.664111 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-log-httpd\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.664195 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-scripts\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.664551 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-run-httpd\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.665032 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.766779 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-run-httpd\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.766892 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.766976 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.767055 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r68z9\" (UniqueName: \"kubernetes.io/projected/4484f510-ece4-42be-b7cd-ede2e47777b1-kube-api-access-r68z9\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.767107 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-config-data\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.767144 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-log-httpd\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.767218 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-scripts\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.767353 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-run-httpd\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.769302 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-log-httpd\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.774899 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.775316 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.784603 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r68z9\" (UniqueName: \"kubernetes.io/projected/4484f510-ece4-42be-b7cd-ede2e47777b1-kube-api-access-r68z9\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.791638 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-scripts\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.792732 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-config-data\") pod \"ceilometer-0\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " pod="openstack/ceilometer-0" Nov 23 07:10:53 crc kubenswrapper[4906]: I1123 07:10:53.872154 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:10:54 crc kubenswrapper[4906]: I1123 07:10:54.407069 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:54 crc kubenswrapper[4906]: I1123 07:10:54.499779 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerStarted","Data":"9b7a2ab54002741e87ab57b649fbeafc758101e6d18a84ac434462fa0c1a02a7"} Nov 23 07:10:54 crc kubenswrapper[4906]: I1123 07:10:54.985342 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:10:55 crc kubenswrapper[4906]: I1123 07:10:55.369770 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00e11c87-d705-4b7a-8771-13d3117e67e7" path="/var/lib/kubelet/pods/00e11c87-d705-4b7a-8771-13d3117e67e7/volumes" Nov 23 07:10:55 crc kubenswrapper[4906]: I1123 07:10:55.370936 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4533768d-d850-457e-87cf-7adf0d594915" path="/var/lib/kubelet/pods/4533768d-d850-457e-87cf-7adf0d594915/volumes" Nov 23 07:10:55 crc kubenswrapper[4906]: I1123 07:10:55.510334 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerStarted","Data":"22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec"} Nov 23 07:10:56 crc kubenswrapper[4906]: I1123 07:10:56.521961 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerStarted","Data":"461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72"} Nov 23 07:10:57 crc kubenswrapper[4906]: I1123 07:10:57.534924 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerStarted","Data":"76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46"} Nov 23 07:10:58 crc kubenswrapper[4906]: I1123 07:10:58.552218 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerStarted","Data":"1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61"} Nov 23 07:10:58 crc kubenswrapper[4906]: I1123 07:10:58.552579 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-central-agent" containerID="cri-o://22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec" gracePeriod=30 Nov 23 07:10:58 crc kubenswrapper[4906]: I1123 07:10:58.552897 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="proxy-httpd" containerID="cri-o://1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61" gracePeriod=30 Nov 23 07:10:58 crc kubenswrapper[4906]: I1123 07:10:58.552954 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="sg-core" containerID="cri-o://76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46" gracePeriod=30 Nov 23 07:10:58 crc kubenswrapper[4906]: I1123 07:10:58.552987 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-notification-agent" containerID="cri-o://461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72" gracePeriod=30 Nov 23 07:10:58 crc kubenswrapper[4906]: I1123 07:10:58.553963 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:10:58 crc kubenswrapper[4906]: I1123 07:10:58.597790 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.314178673 podStartE2EDuration="5.597755932s" podCreationTimestamp="2025-11-23 07:10:53 +0000 UTC" firstStartedPulling="2025-11-23 07:10:54.394606289 +0000 UTC m=+1269.907997592" lastFinishedPulling="2025-11-23 07:10:57.678183518 +0000 UTC m=+1273.191574851" observedRunningTime="2025-11-23 07:10:58.590126743 +0000 UTC m=+1274.103518086" watchObservedRunningTime="2025-11-23 07:10:58.597755932 +0000 UTC m=+1274.111147275" Nov 23 07:10:59 crc kubenswrapper[4906]: I1123 07:10:59.568814 4906 generic.go:334] "Generic (PLEG): container finished" podID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerID="1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61" exitCode=0 Nov 23 07:10:59 crc kubenswrapper[4906]: I1123 07:10:59.568856 4906 generic.go:334] "Generic (PLEG): container finished" podID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerID="76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46" exitCode=2 Nov 23 07:10:59 crc kubenswrapper[4906]: I1123 07:10:59.568866 4906 generic.go:334] "Generic (PLEG): container finished" podID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerID="461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72" exitCode=0 Nov 23 07:10:59 crc kubenswrapper[4906]: I1123 07:10:59.568922 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerDied","Data":"1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61"} Nov 23 07:10:59 crc kubenswrapper[4906]: I1123 07:10:59.569077 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerDied","Data":"76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46"} Nov 23 07:10:59 crc kubenswrapper[4906]: I1123 07:10:59.569112 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerDied","Data":"461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72"} Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.325585 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.443873 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-log-httpd\") pod \"4484f510-ece4-42be-b7cd-ede2e47777b1\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.443973 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-combined-ca-bundle\") pod \"4484f510-ece4-42be-b7cd-ede2e47777b1\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.444013 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-sg-core-conf-yaml\") pod \"4484f510-ece4-42be-b7cd-ede2e47777b1\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.444045 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r68z9\" (UniqueName: \"kubernetes.io/projected/4484f510-ece4-42be-b7cd-ede2e47777b1-kube-api-access-r68z9\") pod \"4484f510-ece4-42be-b7cd-ede2e47777b1\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.444105 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-scripts\") pod \"4484f510-ece4-42be-b7cd-ede2e47777b1\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.444146 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-config-data\") pod \"4484f510-ece4-42be-b7cd-ede2e47777b1\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.444228 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-run-httpd\") pod \"4484f510-ece4-42be-b7cd-ede2e47777b1\" (UID: \"4484f510-ece4-42be-b7cd-ede2e47777b1\") " Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.445029 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4484f510-ece4-42be-b7cd-ede2e47777b1" (UID: "4484f510-ece4-42be-b7cd-ede2e47777b1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.446781 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4484f510-ece4-42be-b7cd-ede2e47777b1" (UID: "4484f510-ece4-42be-b7cd-ede2e47777b1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.453196 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-scripts" (OuterVolumeSpecName: "scripts") pod "4484f510-ece4-42be-b7cd-ede2e47777b1" (UID: "4484f510-ece4-42be-b7cd-ede2e47777b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.453264 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4484f510-ece4-42be-b7cd-ede2e47777b1-kube-api-access-r68z9" (OuterVolumeSpecName: "kube-api-access-r68z9") pod "4484f510-ece4-42be-b7cd-ede2e47777b1" (UID: "4484f510-ece4-42be-b7cd-ede2e47777b1"). InnerVolumeSpecName "kube-api-access-r68z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.479293 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4484f510-ece4-42be-b7cd-ede2e47777b1" (UID: "4484f510-ece4-42be-b7cd-ede2e47777b1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.547432 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.547476 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.547490 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4484f510-ece4-42be-b7cd-ede2e47777b1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.547499 4906 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.547511 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r68z9\" (UniqueName: \"kubernetes.io/projected/4484f510-ece4-42be-b7cd-ede2e47777b1-kube-api-access-r68z9\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.558586 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-config-data" (OuterVolumeSpecName: "config-data") pod "4484f510-ece4-42be-b7cd-ede2e47777b1" (UID: "4484f510-ece4-42be-b7cd-ede2e47777b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.574427 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4484f510-ece4-42be-b7cd-ede2e47777b1" (UID: "4484f510-ece4-42be-b7cd-ede2e47777b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.639667 4906 generic.go:334] "Generic (PLEG): container finished" podID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerID="22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec" exitCode=0 Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.639744 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerDied","Data":"22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec"} Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.639781 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4484f510-ece4-42be-b7cd-ede2e47777b1","Type":"ContainerDied","Data":"9b7a2ab54002741e87ab57b649fbeafc758101e6d18a84ac434462fa0c1a02a7"} Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.639804 4906 scope.go:117] "RemoveContainer" containerID="1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.639893 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.649896 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.649938 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4484f510-ece4-42be-b7cd-ede2e47777b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.662214 4906 scope.go:117] "RemoveContainer" containerID="76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.688614 4906 scope.go:117] "RemoveContainer" containerID="461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.700517 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.714339 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.717822 4906 scope.go:117] "RemoveContainer" containerID="22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.731920 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.732606 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-notification-agent" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732629 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-notification-agent" Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.732642 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-central-agent" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732649 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-central-agent" Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.732673 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="sg-core" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732693 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="sg-core" Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.732704 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="proxy-httpd" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732712 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="proxy-httpd" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732949 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-notification-agent" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732959 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="ceilometer-central-agent" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732966 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="sg-core" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.732977 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" containerName="proxy-httpd" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.736514 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.740807 4906 scope.go:117] "RemoveContainer" containerID="1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61" Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.741833 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61\": container with ID starting with 1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61 not found: ID does not exist" containerID="1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.741885 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61"} err="failed to get container status \"1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61\": rpc error: code = NotFound desc = could not find container \"1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61\": container with ID starting with 1074da0b7b44e26515190516304857542ca73b231c53eb7143d093ac6c125f61 not found: ID does not exist" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.741922 4906 scope.go:117] "RemoveContainer" containerID="76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.742391 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.744367 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46\": container with ID starting with 76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46 not found: ID does not exist" containerID="76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.744420 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46"} err="failed to get container status \"76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46\": rpc error: code = NotFound desc = could not find container \"76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46\": container with ID starting with 76bb0c04d490ee3176a740a4684745b2a68fc7520fbcef0bda37b00b88bdaa46 not found: ID does not exist" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.744443 4906 scope.go:117] "RemoveContainer" containerID="461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.745884 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.747999 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72\": container with ID starting with 461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72 not found: ID does not exist" containerID="461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.748035 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72"} err="failed to get container status \"461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72\": rpc error: code = NotFound desc = could not find container \"461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72\": container with ID starting with 461601a09bd2cf261e369c6b3ca0efe10e5a3464e611088878d6ae4946c5da72 not found: ID does not exist" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.748071 4906 scope.go:117] "RemoveContainer" containerID="22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec" Nov 23 07:11:05 crc kubenswrapper[4906]: E1123 07:11:05.751814 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec\": container with ID starting with 22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec not found: ID does not exist" containerID="22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.751842 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec"} err="failed to get container status \"22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec\": rpc error: code = NotFound desc = could not find container \"22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec\": container with ID starting with 22c1836bcd75233539165427d7c62e29188d073fb2392822510bdf3f0eafe6ec not found: ID does not exist" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.755094 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.756564 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-log-httpd\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.756895 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.757118 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-run-httpd\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.757305 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-scripts\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.757486 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvgtf\" (UniqueName: \"kubernetes.io/projected/65b17bcd-4689-4bb4-aea9-47288ea27f7d-kube-api-access-gvgtf\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.757591 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.757843 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-config-data\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.860050 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-run-httpd\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.860506 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-scripts\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.860653 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvgtf\" (UniqueName: \"kubernetes.io/projected/65b17bcd-4689-4bb4-aea9-47288ea27f7d-kube-api-access-gvgtf\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.860792 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.860933 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-config-data\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.860801 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-run-httpd\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.861121 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-log-httpd\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.861367 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.861751 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-log-httpd\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.865820 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.866153 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.866671 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-scripts\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.867100 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-config-data\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:05 crc kubenswrapper[4906]: I1123 07:11:05.882493 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvgtf\" (UniqueName: \"kubernetes.io/projected/65b17bcd-4689-4bb4-aea9-47288ea27f7d-kube-api-access-gvgtf\") pod \"ceilometer-0\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " pod="openstack/ceilometer-0" Nov 23 07:11:06 crc kubenswrapper[4906]: I1123 07:11:06.062957 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:06 crc kubenswrapper[4906]: I1123 07:11:06.540107 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:06 crc kubenswrapper[4906]: I1123 07:11:06.652957 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerStarted","Data":"679d698ff9ded65f15d5398d00a28069935d9fbe5d69037c33badcd4c9078dce"} Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.142717 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4r2qz"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.146908 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.186991 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78dd4ca-5705-4554-bdc0-1ec212b5751e-operator-scripts\") pod \"nova-api-db-create-4r2qz\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.187044 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94qhz\" (UniqueName: \"kubernetes.io/projected/e78dd4ca-5705-4554-bdc0-1ec212b5751e-kube-api-access-94qhz\") pod \"nova-api-db-create-4r2qz\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.200772 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4r2qz"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.261995 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-x42nt"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.263616 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.273102 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x42nt"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.289539 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsz64\" (UniqueName: \"kubernetes.io/projected/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-kube-api-access-bsz64\") pod \"nova-cell0-db-create-x42nt\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.289798 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-operator-scripts\") pod \"nova-cell0-db-create-x42nt\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.289978 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78dd4ca-5705-4554-bdc0-1ec212b5751e-operator-scripts\") pod \"nova-api-db-create-4r2qz\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.290068 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94qhz\" (UniqueName: \"kubernetes.io/projected/e78dd4ca-5705-4554-bdc0-1ec212b5751e-kube-api-access-94qhz\") pod \"nova-api-db-create-4r2qz\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.291886 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78dd4ca-5705-4554-bdc0-1ec212b5751e-operator-scripts\") pod \"nova-api-db-create-4r2qz\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.317042 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94qhz\" (UniqueName: \"kubernetes.io/projected/e78dd4ca-5705-4554-bdc0-1ec212b5751e-kube-api-access-94qhz\") pod \"nova-api-db-create-4r2qz\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.355656 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-944e-account-create-zh9ms"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.360071 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.363053 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.373981 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4484f510-ece4-42be-b7cd-ede2e47777b1" path="/var/lib/kubelet/pods/4484f510-ece4-42be-b7cd-ede2e47777b1/volumes" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.374783 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-944e-account-create-zh9ms"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.403364 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsz64\" (UniqueName: \"kubernetes.io/projected/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-kube-api-access-bsz64\") pod \"nova-cell0-db-create-x42nt\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.403461 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-operator-scripts\") pod \"nova-cell0-db-create-x42nt\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.403508 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40cb3343-177c-42cf-8604-c22eaca5e593-operator-scripts\") pod \"nova-api-944e-account-create-zh9ms\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.403706 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7qw5\" (UniqueName: \"kubernetes.io/projected/40cb3343-177c-42cf-8604-c22eaca5e593-kube-api-access-p7qw5\") pod \"nova-api-944e-account-create-zh9ms\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.406275 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-operator-scripts\") pod \"nova-cell0-db-create-x42nt\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.438925 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsz64\" (UniqueName: \"kubernetes.io/projected/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-kube-api-access-bsz64\") pod \"nova-cell0-db-create-x42nt\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.467749 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-wbdjb"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.471012 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.479995 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wbdjb"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.505707 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40cb3343-177c-42cf-8604-c22eaca5e593-operator-scripts\") pod \"nova-api-944e-account-create-zh9ms\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.505779 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68lrp\" (UniqueName: \"kubernetes.io/projected/fc92acb2-4da0-4ff3-8489-e63fb755861b-kube-api-access-68lrp\") pod \"nova-cell1-db-create-wbdjb\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.505823 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc92acb2-4da0-4ff3-8489-e63fb755861b-operator-scripts\") pod \"nova-cell1-db-create-wbdjb\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.505860 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7qw5\" (UniqueName: \"kubernetes.io/projected/40cb3343-177c-42cf-8604-c22eaca5e593-kube-api-access-p7qw5\") pod \"nova-api-944e-account-create-zh9ms\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.506606 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40cb3343-177c-42cf-8604-c22eaca5e593-operator-scripts\") pod \"nova-api-944e-account-create-zh9ms\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.515261 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.528214 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7qw5\" (UniqueName: \"kubernetes.io/projected/40cb3343-177c-42cf-8604-c22eaca5e593-kube-api-access-p7qw5\") pod \"nova-api-944e-account-create-zh9ms\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.557769 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-82ff-account-create-5xzpr"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.559222 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.561632 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.563446 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-82ff-account-create-5xzpr"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.594792 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.609972 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nxmq\" (UniqueName: \"kubernetes.io/projected/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-kube-api-access-2nxmq\") pod \"nova-cell0-82ff-account-create-5xzpr\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.610022 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-operator-scripts\") pod \"nova-cell0-82ff-account-create-5xzpr\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.610080 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68lrp\" (UniqueName: \"kubernetes.io/projected/fc92acb2-4da0-4ff3-8489-e63fb755861b-kube-api-access-68lrp\") pod \"nova-cell1-db-create-wbdjb\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.610114 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc92acb2-4da0-4ff3-8489-e63fb755861b-operator-scripts\") pod \"nova-cell1-db-create-wbdjb\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.628248 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc92acb2-4da0-4ff3-8489-e63fb755861b-operator-scripts\") pod \"nova-cell1-db-create-wbdjb\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.649239 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68lrp\" (UniqueName: \"kubernetes.io/projected/fc92acb2-4da0-4ff3-8489-e63fb755861b-kube-api-access-68lrp\") pod \"nova-cell1-db-create-wbdjb\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.691490 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.712163 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nxmq\" (UniqueName: \"kubernetes.io/projected/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-kube-api-access-2nxmq\") pod \"nova-cell0-82ff-account-create-5xzpr\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.712230 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-operator-scripts\") pod \"nova-cell0-82ff-account-create-5xzpr\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.713565 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-operator-scripts\") pod \"nova-cell0-82ff-account-create-5xzpr\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.714502 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerStarted","Data":"47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd"} Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.781910 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nxmq\" (UniqueName: \"kubernetes.io/projected/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-kube-api-access-2nxmq\") pod \"nova-cell0-82ff-account-create-5xzpr\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.800463 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.878270 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.881073 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-3763-account-create-stq74"] Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.904269 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.917786 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 23 07:11:07 crc kubenswrapper[4906]: I1123 07:11:07.937874 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3763-account-create-stq74"] Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.049561 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-operator-scripts\") pod \"nova-cell1-3763-account-create-stq74\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.050488 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8phc\" (UniqueName: \"kubernetes.io/projected/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-kube-api-access-d8phc\") pod \"nova-cell1-3763-account-create-stq74\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.152065 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-operator-scripts\") pod \"nova-cell1-3763-account-create-stq74\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.152156 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8phc\" (UniqueName: \"kubernetes.io/projected/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-kube-api-access-d8phc\") pod \"nova-cell1-3763-account-create-stq74\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.153199 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-operator-scripts\") pod \"nova-cell1-3763-account-create-stq74\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.174296 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8phc\" (UniqueName: \"kubernetes.io/projected/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-kube-api-access-d8phc\") pod \"nova-cell1-3763-account-create-stq74\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.281764 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.311193 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4r2qz"] Nov 23 07:11:08 crc kubenswrapper[4906]: W1123 07:11:08.331419 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode78dd4ca_5705_4554_bdc0_1ec212b5751e.slice/crio-df7ac23ee1491ec575a8ad8c2e7640ab4c84c0bc34e5560d39ecfd06770ffd98 WatchSource:0}: Error finding container df7ac23ee1491ec575a8ad8c2e7640ab4c84c0bc34e5560d39ecfd06770ffd98: Status 404 returned error can't find the container with id df7ac23ee1491ec575a8ad8c2e7640ab4c84c0bc34e5560d39ecfd06770ffd98 Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.497138 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-x42nt"] Nov 23 07:11:08 crc kubenswrapper[4906]: W1123 07:11:08.506922 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ed64d45_85e5_4a97_beb7_8c31e2a91ca2.slice/crio-9a1baad473543a4a166d26550b43bed2e64174d991985d1b59188866eea870fd WatchSource:0}: Error finding container 9a1baad473543a4a166d26550b43bed2e64174d991985d1b59188866eea870fd: Status 404 returned error can't find the container with id 9a1baad473543a4a166d26550b43bed2e64174d991985d1b59188866eea870fd Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.609491 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-944e-account-create-zh9ms"] Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.629869 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wbdjb"] Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.639288 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-82ff-account-create-5xzpr"] Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.731827 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-82ff-account-create-5xzpr" event={"ID":"08d0ca8b-c938-4cdb-8a50-223295bcc5e4","Type":"ContainerStarted","Data":"29e8095c68935786a59b57bc2715def1c70fbef5a940dc17e6f249f95c1b24fa"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.733497 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x42nt" event={"ID":"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2","Type":"ContainerStarted","Data":"9fd4e4ce70e7b4e8a8a0cf7929ef4a4d7fd0ebe705d7489ea597f1eedf4b7d4e"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.733520 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x42nt" event={"ID":"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2","Type":"ContainerStarted","Data":"9a1baad473543a4a166d26550b43bed2e64174d991985d1b59188866eea870fd"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.742076 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4r2qz" event={"ID":"e78dd4ca-5705-4554-bdc0-1ec212b5751e","Type":"ContainerStarted","Data":"dd3f69e2e62458ca440ee946a42dded4d158d99cfba6eb855675caa7f5eb42e5"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.742133 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4r2qz" event={"ID":"e78dd4ca-5705-4554-bdc0-1ec212b5751e","Type":"ContainerStarted","Data":"df7ac23ee1491ec575a8ad8c2e7640ab4c84c0bc34e5560d39ecfd06770ffd98"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.747815 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wbdjb" event={"ID":"fc92acb2-4da0-4ff3-8489-e63fb755861b","Type":"ContainerStarted","Data":"a9cbc58ead6f94229aeb20eb1a03f1302a0d4c69c683d92ff6248eb20507065d"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.752691 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerStarted","Data":"3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.754942 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-944e-account-create-zh9ms" event={"ID":"40cb3343-177c-42cf-8604-c22eaca5e593","Type":"ContainerStarted","Data":"3966cd3091f07574daf7a4a2dbc25195337429b7caf4d54ad90e42b3aac82ab8"} Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.756596 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-x42nt" podStartSLOduration=1.75657366 podStartE2EDuration="1.75657366s" podCreationTimestamp="2025-11-23 07:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:11:08.748189551 +0000 UTC m=+1284.261580854" watchObservedRunningTime="2025-11-23 07:11:08.75657366 +0000 UTC m=+1284.269964963" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.770998 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-4r2qz" podStartSLOduration=1.770975957 podStartE2EDuration="1.770975957s" podCreationTimestamp="2025-11-23 07:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:11:08.760925884 +0000 UTC m=+1284.274317187" watchObservedRunningTime="2025-11-23 07:11:08.770975957 +0000 UTC m=+1284.284367260" Nov 23 07:11:08 crc kubenswrapper[4906]: I1123 07:11:08.848420 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3763-account-create-stq74"] Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.766998 4906 generic.go:334] "Generic (PLEG): container finished" podID="e78dd4ca-5705-4554-bdc0-1ec212b5751e" containerID="dd3f69e2e62458ca440ee946a42dded4d158d99cfba6eb855675caa7f5eb42e5" exitCode=0 Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.767202 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4r2qz" event={"ID":"e78dd4ca-5705-4554-bdc0-1ec212b5751e","Type":"ContainerDied","Data":"dd3f69e2e62458ca440ee946a42dded4d158d99cfba6eb855675caa7f5eb42e5"} Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.780755 4906 generic.go:334] "Generic (PLEG): container finished" podID="03f12dad-1cb2-4e43-b9cc-7658ed44bc1b" containerID="6a774e7bd820a8715662dc5a9ea4bf2e91f3dcc4e667d067b8591e6ae75de44b" exitCode=0 Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.780971 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3763-account-create-stq74" event={"ID":"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b","Type":"ContainerDied","Data":"6a774e7bd820a8715662dc5a9ea4bf2e91f3dcc4e667d067b8591e6ae75de44b"} Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.781044 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3763-account-create-stq74" event={"ID":"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b","Type":"ContainerStarted","Data":"e5648dbf20e5623ded1c7f1fa94a9d7b4304d96a74ee1a9ce3ceab447fd38c03"} Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.791070 4906 generic.go:334] "Generic (PLEG): container finished" podID="fc92acb2-4da0-4ff3-8489-e63fb755861b" containerID="6ba149ba0e302725b1fbc3ed83081a7cae1d95f6be1c61dedefc243562bcc3ec" exitCode=0 Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.791172 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wbdjb" event={"ID":"fc92acb2-4da0-4ff3-8489-e63fb755861b","Type":"ContainerDied","Data":"6ba149ba0e302725b1fbc3ed83081a7cae1d95f6be1c61dedefc243562bcc3ec"} Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.800919 4906 generic.go:334] "Generic (PLEG): container finished" podID="40cb3343-177c-42cf-8604-c22eaca5e593" containerID="659ee25bbd94087f3829983a3b02a6c010a9ff2d2e93cdd5a42d2537a3e03b42" exitCode=0 Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.801077 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-944e-account-create-zh9ms" event={"ID":"40cb3343-177c-42cf-8604-c22eaca5e593","Type":"ContainerDied","Data":"659ee25bbd94087f3829983a3b02a6c010a9ff2d2e93cdd5a42d2537a3e03b42"} Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.807939 4906 generic.go:334] "Generic (PLEG): container finished" podID="08d0ca8b-c938-4cdb-8a50-223295bcc5e4" containerID="fb2db4868429914c4974cfcbe6924d07ccd95b04a3c7bb744e5c5081ac478bad" exitCode=0 Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.808057 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-82ff-account-create-5xzpr" event={"ID":"08d0ca8b-c938-4cdb-8a50-223295bcc5e4","Type":"ContainerDied","Data":"fb2db4868429914c4974cfcbe6924d07ccd95b04a3c7bb744e5c5081ac478bad"} Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.824628 4906 generic.go:334] "Generic (PLEG): container finished" podID="7ed64d45-85e5-4a97-beb7-8c31e2a91ca2" containerID="9fd4e4ce70e7b4e8a8a0cf7929ef4a4d7fd0ebe705d7489ea597f1eedf4b7d4e" exitCode=0 Nov 23 07:11:09 crc kubenswrapper[4906]: I1123 07:11:09.824731 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x42nt" event={"ID":"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2","Type":"ContainerDied","Data":"9fd4e4ce70e7b4e8a8a0cf7929ef4a4d7fd0ebe705d7489ea597f1eedf4b7d4e"} Nov 23 07:11:10 crc kubenswrapper[4906]: I1123 07:11:10.841876 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerStarted","Data":"9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.285712 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.347544 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40cb3343-177c-42cf-8604-c22eaca5e593-operator-scripts\") pod \"40cb3343-177c-42cf-8604-c22eaca5e593\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.347586 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7qw5\" (UniqueName: \"kubernetes.io/projected/40cb3343-177c-42cf-8604-c22eaca5e593-kube-api-access-p7qw5\") pod \"40cb3343-177c-42cf-8604-c22eaca5e593\" (UID: \"40cb3343-177c-42cf-8604-c22eaca5e593\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.348323 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40cb3343-177c-42cf-8604-c22eaca5e593-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40cb3343-177c-42cf-8604-c22eaca5e593" (UID: "40cb3343-177c-42cf-8604-c22eaca5e593"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.358570 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40cb3343-177c-42cf-8604-c22eaca5e593-kube-api-access-p7qw5" (OuterVolumeSpecName: "kube-api-access-p7qw5") pod "40cb3343-177c-42cf-8604-c22eaca5e593" (UID: "40cb3343-177c-42cf-8604-c22eaca5e593"). InnerVolumeSpecName "kube-api-access-p7qw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.450601 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40cb3343-177c-42cf-8604-c22eaca5e593-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.450629 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7qw5\" (UniqueName: \"kubernetes.io/projected/40cb3343-177c-42cf-8604-c22eaca5e593-kube-api-access-p7qw5\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.508428 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.529841 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.555841 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.556951 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-log" containerID="cri-o://a7efd54a95aafa3d3cc917715d3b8d6474962d24f6d6688799722b7635d6e399" gracePeriod=30 Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.557055 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-httpd" containerID="cri-o://dae906aff5a2080ffb6904c4f1c75020ff73527712718bb4c1180d689e9bae6d" gracePeriod=30 Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.561957 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.582960 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.584876 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654303 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8phc\" (UniqueName: \"kubernetes.io/projected/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-kube-api-access-d8phc\") pod \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654483 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-operator-scripts\") pod \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654564 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nxmq\" (UniqueName: \"kubernetes.io/projected/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-kube-api-access-2nxmq\") pod \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\" (UID: \"08d0ca8b-c938-4cdb-8a50-223295bcc5e4\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654606 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94qhz\" (UniqueName: \"kubernetes.io/projected/e78dd4ca-5705-4554-bdc0-1ec212b5751e-kube-api-access-94qhz\") pod \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654629 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68lrp\" (UniqueName: \"kubernetes.io/projected/fc92acb2-4da0-4ff3-8489-e63fb755861b-kube-api-access-68lrp\") pod \"fc92acb2-4da0-4ff3-8489-e63fb755861b\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654718 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78dd4ca-5705-4554-bdc0-1ec212b5751e-operator-scripts\") pod \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\" (UID: \"e78dd4ca-5705-4554-bdc0-1ec212b5751e\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654777 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc92acb2-4da0-4ff3-8489-e63fb755861b-operator-scripts\") pod \"fc92acb2-4da0-4ff3-8489-e63fb755861b\" (UID: \"fc92acb2-4da0-4ff3-8489-e63fb755861b\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654809 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-operator-scripts\") pod \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654861 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsz64\" (UniqueName: \"kubernetes.io/projected/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-kube-api-access-bsz64\") pod \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\" (UID: \"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.654894 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-operator-scripts\") pod \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\" (UID: \"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b\") " Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.656534 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78dd4ca-5705-4554-bdc0-1ec212b5751e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e78dd4ca-5705-4554-bdc0-1ec212b5751e" (UID: "e78dd4ca-5705-4554-bdc0-1ec212b5751e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.656980 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "08d0ca8b-c938-4cdb-8a50-223295bcc5e4" (UID: "08d0ca8b-c938-4cdb-8a50-223295bcc5e4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.657333 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ed64d45-85e5-4a97-beb7-8c31e2a91ca2" (UID: "7ed64d45-85e5-4a97-beb7-8c31e2a91ca2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.658437 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc92acb2-4da0-4ff3-8489-e63fb755861b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc92acb2-4da0-4ff3-8489-e63fb755861b" (UID: "fc92acb2-4da0-4ff3-8489-e63fb755861b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.658421 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "03f12dad-1cb2-4e43-b9cc-7658ed44bc1b" (UID: "03f12dad-1cb2-4e43-b9cc-7658ed44bc1b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.662610 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-kube-api-access-d8phc" (OuterVolumeSpecName: "kube-api-access-d8phc") pod "03f12dad-1cb2-4e43-b9cc-7658ed44bc1b" (UID: "03f12dad-1cb2-4e43-b9cc-7658ed44bc1b"). InnerVolumeSpecName "kube-api-access-d8phc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.663461 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-kube-api-access-bsz64" (OuterVolumeSpecName: "kube-api-access-bsz64") pod "7ed64d45-85e5-4a97-beb7-8c31e2a91ca2" (UID: "7ed64d45-85e5-4a97-beb7-8c31e2a91ca2"). InnerVolumeSpecName "kube-api-access-bsz64". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.663881 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78dd4ca-5705-4554-bdc0-1ec212b5751e-kube-api-access-94qhz" (OuterVolumeSpecName: "kube-api-access-94qhz") pod "e78dd4ca-5705-4554-bdc0-1ec212b5751e" (UID: "e78dd4ca-5705-4554-bdc0-1ec212b5751e"). InnerVolumeSpecName "kube-api-access-94qhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.665366 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-kube-api-access-2nxmq" (OuterVolumeSpecName: "kube-api-access-2nxmq") pod "08d0ca8b-c938-4cdb-8a50-223295bcc5e4" (UID: "08d0ca8b-c938-4cdb-8a50-223295bcc5e4"). InnerVolumeSpecName "kube-api-access-2nxmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.686443 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc92acb2-4da0-4ff3-8489-e63fb755861b-kube-api-access-68lrp" (OuterVolumeSpecName: "kube-api-access-68lrp") pod "fc92acb2-4da0-4ff3-8489-e63fb755861b" (UID: "fc92acb2-4da0-4ff3-8489-e63fb755861b"). InnerVolumeSpecName "kube-api-access-68lrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757363 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc92acb2-4da0-4ff3-8489-e63fb755861b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757405 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757420 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsz64\" (UniqueName: \"kubernetes.io/projected/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2-kube-api-access-bsz64\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757434 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757443 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8phc\" (UniqueName: \"kubernetes.io/projected/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b-kube-api-access-d8phc\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757451 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757461 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nxmq\" (UniqueName: \"kubernetes.io/projected/08d0ca8b-c938-4cdb-8a50-223295bcc5e4-kube-api-access-2nxmq\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757469 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94qhz\" (UniqueName: \"kubernetes.io/projected/e78dd4ca-5705-4554-bdc0-1ec212b5751e-kube-api-access-94qhz\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757478 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68lrp\" (UniqueName: \"kubernetes.io/projected/fc92acb2-4da0-4ff3-8489-e63fb755861b-kube-api-access-68lrp\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.757488 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78dd4ca-5705-4554-bdc0-1ec212b5751e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.852452 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-82ff-account-create-5xzpr" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.852902 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-82ff-account-create-5xzpr" event={"ID":"08d0ca8b-c938-4cdb-8a50-223295bcc5e4","Type":"ContainerDied","Data":"29e8095c68935786a59b57bc2715def1c70fbef5a940dc17e6f249f95c1b24fa"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.852952 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29e8095c68935786a59b57bc2715def1c70fbef5a940dc17e6f249f95c1b24fa" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.854532 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-x42nt" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.854538 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-x42nt" event={"ID":"7ed64d45-85e5-4a97-beb7-8c31e2a91ca2","Type":"ContainerDied","Data":"9a1baad473543a4a166d26550b43bed2e64174d991985d1b59188866eea870fd"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.854734 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a1baad473543a4a166d26550b43bed2e64174d991985d1b59188866eea870fd" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.855813 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4r2qz" event={"ID":"e78dd4ca-5705-4554-bdc0-1ec212b5751e","Type":"ContainerDied","Data":"df7ac23ee1491ec575a8ad8c2e7640ab4c84c0bc34e5560d39ecfd06770ffd98"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.855828 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4r2qz" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.855839 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df7ac23ee1491ec575a8ad8c2e7640ab4c84c0bc34e5560d39ecfd06770ffd98" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.864563 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3763-account-create-stq74" event={"ID":"03f12dad-1cb2-4e43-b9cc-7658ed44bc1b","Type":"ContainerDied","Data":"e5648dbf20e5623ded1c7f1fa94a9d7b4304d96a74ee1a9ce3ceab447fd38c03"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.864593 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5648dbf20e5623ded1c7f1fa94a9d7b4304d96a74ee1a9ce3ceab447fd38c03" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.864639 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3763-account-create-stq74" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.866979 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wbdjb" event={"ID":"fc92acb2-4da0-4ff3-8489-e63fb755861b","Type":"ContainerDied","Data":"a9cbc58ead6f94229aeb20eb1a03f1302a0d4c69c683d92ff6248eb20507065d"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.867007 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9cbc58ead6f94229aeb20eb1a03f1302a0d4c69c683d92ff6248eb20507065d" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.867097 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wbdjb" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.869685 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerStarted","Data":"d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.869744 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.872022 4906 generic.go:334] "Generic (PLEG): container finished" podID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerID="a7efd54a95aafa3d3cc917715d3b8d6474962d24f6d6688799722b7635d6e399" exitCode=143 Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.872084 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e6209532-4ce6-482f-9c69-021b4ec0b682","Type":"ContainerDied","Data":"a7efd54a95aafa3d3cc917715d3b8d6474962d24f6d6688799722b7635d6e399"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.874746 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-944e-account-create-zh9ms" event={"ID":"40cb3343-177c-42cf-8604-c22eaca5e593","Type":"ContainerDied","Data":"3966cd3091f07574daf7a4a2dbc25195337429b7caf4d54ad90e42b3aac82ab8"} Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.874773 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3966cd3091f07574daf7a4a2dbc25195337429b7caf4d54ad90e42b3aac82ab8" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.874849 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-944e-account-create-zh9ms" Nov 23 07:11:11 crc kubenswrapper[4906]: I1123 07:11:11.895660 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.37960562 podStartE2EDuration="6.895635262s" podCreationTimestamp="2025-11-23 07:11:05 +0000 UTC" firstStartedPulling="2025-11-23 07:11:06.550762602 +0000 UTC m=+1282.064153905" lastFinishedPulling="2025-11-23 07:11:11.066792244 +0000 UTC m=+1286.580183547" observedRunningTime="2025-11-23 07:11:11.892754987 +0000 UTC m=+1287.406146290" watchObservedRunningTime="2025-11-23 07:11:11.895635262 +0000 UTC m=+1287.409026575" Nov 23 07:11:12 crc kubenswrapper[4906]: I1123 07:11:12.628824 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:11:12 crc kubenswrapper[4906]: I1123 07:11:12.629432 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-log" containerID="cri-o://2b9c7e90e2950b3fcc525bdf5a7dbb42733038fec6d65bd737ad2199a9071b46" gracePeriod=30 Nov 23 07:11:12 crc kubenswrapper[4906]: I1123 07:11:12.629593 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-httpd" containerID="cri-o://20b05ca7f58640a43e4cde5f3c87862d66748a5431dc1655c260c2a29bf3dae2" gracePeriod=30 Nov 23 07:11:12 crc kubenswrapper[4906]: I1123 07:11:12.890817 4906 generic.go:334] "Generic (PLEG): container finished" podID="9e594b90-2851-4374-8913-08103bb5065e" containerID="2b9c7e90e2950b3fcc525bdf5a7dbb42733038fec6d65bd737ad2199a9071b46" exitCode=143 Nov 23 07:11:12 crc kubenswrapper[4906]: I1123 07:11:12.891346 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9e594b90-2851-4374-8913-08103bb5065e","Type":"ContainerDied","Data":"2b9c7e90e2950b3fcc525bdf5a7dbb42733038fec6d65bd737ad2199a9071b46"} Nov 23 07:11:13 crc kubenswrapper[4906]: I1123 07:11:13.736036 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:13 crc kubenswrapper[4906]: I1123 07:11:13.899039 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-central-agent" containerID="cri-o://47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd" gracePeriod=30 Nov 23 07:11:13 crc kubenswrapper[4906]: I1123 07:11:13.899102 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="sg-core" containerID="cri-o://9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551" gracePeriod=30 Nov 23 07:11:13 crc kubenswrapper[4906]: I1123 07:11:13.899123 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-notification-agent" containerID="cri-o://3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3" gracePeriod=30 Nov 23 07:11:13 crc kubenswrapper[4906]: I1123 07:11:13.899231 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="proxy-httpd" containerID="cri-o://d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87" gracePeriod=30 Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.917561 4906 generic.go:334] "Generic (PLEG): container finished" podID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerID="dae906aff5a2080ffb6904c4f1c75020ff73527712718bb4c1180d689e9bae6d" exitCode=0 Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.918154 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e6209532-4ce6-482f-9c69-021b4ec0b682","Type":"ContainerDied","Data":"dae906aff5a2080ffb6904c4f1c75020ff73527712718bb4c1180d689e9bae6d"} Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.922939 4906 generic.go:334] "Generic (PLEG): container finished" podID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerID="d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87" exitCode=0 Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.922964 4906 generic.go:334] "Generic (PLEG): container finished" podID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerID="9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551" exitCode=2 Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.922972 4906 generic.go:334] "Generic (PLEG): container finished" podID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerID="3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3" exitCode=0 Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.922990 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerDied","Data":"d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87"} Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.923009 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerDied","Data":"9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551"} Nov 23 07:11:14 crc kubenswrapper[4906]: I1123 07:11:14.923019 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerDied","Data":"3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3"} Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.375512 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444352 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-combined-ca-bundle\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444443 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-scripts\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444483 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444508 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh4pc\" (UniqueName: \"kubernetes.io/projected/e6209532-4ce6-482f-9c69-021b4ec0b682-kube-api-access-wh4pc\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444623 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-config-data\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444658 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-public-tls-certs\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444799 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-httpd-run\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.444860 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-logs\") pod \"e6209532-4ce6-482f-9c69-021b4ec0b682\" (UID: \"e6209532-4ce6-482f-9c69-021b4ec0b682\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.445715 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-logs" (OuterVolumeSpecName: "logs") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.448554 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.457321 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-scripts" (OuterVolumeSpecName: "scripts") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.457650 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.458337 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6209532-4ce6-482f-9c69-021b4ec0b682-kube-api-access-wh4pc" (OuterVolumeSpecName: "kube-api-access-wh4pc") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "kube-api-access-wh4pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.506397 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-config-data" (OuterVolumeSpecName: "config-data") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.539778 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.541357 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e6209532-4ce6-482f-9c69-021b4ec0b682" (UID: "e6209532-4ce6-482f-9c69-021b4ec0b682"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547670 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547726 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547743 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547755 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6209532-4ce6-482f-9c69-021b4ec0b682-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547770 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547781 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6209532-4ce6-482f-9c69-021b4ec0b682-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547828 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.547841 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh4pc\" (UniqueName: \"kubernetes.io/projected/e6209532-4ce6-482f-9c69-021b4ec0b682-kube-api-access-wh4pc\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.570567 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.588069 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.650270 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.752064 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-config-data\") pod \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.752776 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-run-httpd\") pod \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.752802 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-combined-ca-bundle\") pod \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.753661 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "65b17bcd-4689-4bb4-aea9-47288ea27f7d" (UID: "65b17bcd-4689-4bb4-aea9-47288ea27f7d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.753962 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvgtf\" (UniqueName: \"kubernetes.io/projected/65b17bcd-4689-4bb4-aea9-47288ea27f7d-kube-api-access-gvgtf\") pod \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.754070 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-scripts\") pod \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.754145 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-log-httpd\") pod \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.754216 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-sg-core-conf-yaml\") pod \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\" (UID: \"65b17bcd-4689-4bb4-aea9-47288ea27f7d\") " Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.755146 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "65b17bcd-4689-4bb4-aea9-47288ea27f7d" (UID: "65b17bcd-4689-4bb4-aea9-47288ea27f7d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.756536 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.760362 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b17bcd-4689-4bb4-aea9-47288ea27f7d-kube-api-access-gvgtf" (OuterVolumeSpecName: "kube-api-access-gvgtf") pod "65b17bcd-4689-4bb4-aea9-47288ea27f7d" (UID: "65b17bcd-4689-4bb4-aea9-47288ea27f7d"). InnerVolumeSpecName "kube-api-access-gvgtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.765190 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-scripts" (OuterVolumeSpecName: "scripts") pod "65b17bcd-4689-4bb4-aea9-47288ea27f7d" (UID: "65b17bcd-4689-4bb4-aea9-47288ea27f7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.798215 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "65b17bcd-4689-4bb4-aea9-47288ea27f7d" (UID: "65b17bcd-4689-4bb4-aea9-47288ea27f7d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.864193 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvgtf\" (UniqueName: \"kubernetes.io/projected/65b17bcd-4689-4bb4-aea9-47288ea27f7d-kube-api-access-gvgtf\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.864243 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.864254 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b17bcd-4689-4bb4-aea9-47288ea27f7d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.864265 4906 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.872772 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65b17bcd-4689-4bb4-aea9-47288ea27f7d" (UID: "65b17bcd-4689-4bb4-aea9-47288ea27f7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.899123 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-config-data" (OuterVolumeSpecName: "config-data") pod "65b17bcd-4689-4bb4-aea9-47288ea27f7d" (UID: "65b17bcd-4689-4bb4-aea9-47288ea27f7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.936547 4906 generic.go:334] "Generic (PLEG): container finished" podID="9e594b90-2851-4374-8913-08103bb5065e" containerID="20b05ca7f58640a43e4cde5f3c87862d66748a5431dc1655c260c2a29bf3dae2" exitCode=0 Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.936630 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9e594b90-2851-4374-8913-08103bb5065e","Type":"ContainerDied","Data":"20b05ca7f58640a43e4cde5f3c87862d66748a5431dc1655c260c2a29bf3dae2"} Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.939447 4906 generic.go:334] "Generic (PLEG): container finished" podID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerID="47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd" exitCode=0 Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.939509 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerDied","Data":"47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd"} Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.939543 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b17bcd-4689-4bb4-aea9-47288ea27f7d","Type":"ContainerDied","Data":"679d698ff9ded65f15d5398d00a28069935d9fbe5d69037c33badcd4c9078dce"} Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.939566 4906 scope.go:117] "RemoveContainer" containerID="d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.939777 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.945288 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e6209532-4ce6-482f-9c69-021b4ec0b682","Type":"ContainerDied","Data":"28c48540a505a37e47f0494c78d96f8d1c56492f04c126483447b18c686c854b"} Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.945408 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.966024 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.966055 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b17bcd-4689-4bb4-aea9-47288ea27f7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.967113 4906 scope.go:117] "RemoveContainer" containerID="9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551" Nov 23 07:11:15 crc kubenswrapper[4906]: I1123 07:11:15.996413 4906 scope.go:117] "RemoveContainer" containerID="3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.007814 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.025800 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.036353 4906 scope.go:117] "RemoveContainer" containerID="47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.036514 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.052325 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.084284 4906 scope.go:117] "RemoveContainer" containerID="d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.084396 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086024 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-httpd" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086051 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-httpd" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086068 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="sg-core" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086075 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="sg-core" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086085 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40cb3343-177c-42cf-8604-c22eaca5e593" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086091 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="40cb3343-177c-42cf-8604-c22eaca5e593" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086102 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-log" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086109 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-log" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086123 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed64d45-85e5-4a97-beb7-8c31e2a91ca2" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086129 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed64d45-85e5-4a97-beb7-8c31e2a91ca2" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086139 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78dd4ca-5705-4554-bdc0-1ec212b5751e" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086147 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78dd4ca-5705-4554-bdc0-1ec212b5751e" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086163 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-central-agent" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086168 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-central-agent" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086183 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d0ca8b-c938-4cdb-8a50-223295bcc5e4" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086189 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d0ca8b-c938-4cdb-8a50-223295bcc5e4" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086232 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="proxy-httpd" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086239 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="proxy-httpd" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086250 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc92acb2-4da0-4ff3-8489-e63fb755861b" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086257 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc92acb2-4da0-4ff3-8489-e63fb755861b" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086268 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03f12dad-1cb2-4e43-b9cc-7658ed44bc1b" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086274 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03f12dad-1cb2-4e43-b9cc-7658ed44bc1b" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.086284 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-notification-agent" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086311 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-notification-agent" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086492 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78dd4ca-5705-4554-bdc0-1ec212b5751e" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086501 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed64d45-85e5-4a97-beb7-8c31e2a91ca2" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086511 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="proxy-httpd" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086521 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="40cb3343-177c-42cf-8604-c22eaca5e593" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086534 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-log" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086543 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-central-agent" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086556 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" containerName="glance-httpd" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086565 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="03f12dad-1cb2-4e43-b9cc-7658ed44bc1b" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086575 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="ceilometer-notification-agent" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086585 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" containerName="sg-core" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.086594 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc92acb2-4da0-4ff3-8489-e63fb755861b" containerName="mariadb-database-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.087844 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87\": container with ID starting with d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87 not found: ID does not exist" containerID="d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.087883 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87"} err="failed to get container status \"d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87\": rpc error: code = NotFound desc = could not find container \"d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87\": container with ID starting with d9589972c178296e0633ed340daf766d58683484ad20617644f7799bfae24a87 not found: ID does not exist" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.087910 4906 scope.go:117] "RemoveContainer" containerID="9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.088407 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d0ca8b-c938-4cdb-8a50-223295bcc5e4" containerName="mariadb-account-create" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.089179 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551\": container with ID starting with 9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551 not found: ID does not exist" containerID="9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.089205 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551"} err="failed to get container status \"9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551\": rpc error: code = NotFound desc = could not find container \"9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551\": container with ID starting with 9ba231f06fff833d51465d685e759c3fae40087143a431d1122c40ba6f83b551 not found: ID does not exist" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.089221 4906 scope.go:117] "RemoveContainer" containerID="3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.089433 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3\": container with ID starting with 3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3 not found: ID does not exist" containerID="3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.089458 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3"} err="failed to get container status \"3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3\": rpc error: code = NotFound desc = could not find container \"3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3\": container with ID starting with 3a21ca4aed6bc7f6ab8d81956f266606c109c2bbcfd6704fb9043b58493197a3 not found: ID does not exist" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.089472 4906 scope.go:117] "RemoveContainer" containerID="47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd" Nov 23 07:11:16 crc kubenswrapper[4906]: E1123 07:11:16.089748 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd\": container with ID starting with 47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd not found: ID does not exist" containerID="47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.089772 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd"} err="failed to get container status \"47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd\": rpc error: code = NotFound desc = could not find container \"47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd\": container with ID starting with 47afc05530a2cc4565b17b57a57ae5e524d235ea733443dc5dec9ffbd03e1bcd not found: ID does not exist" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.089786 4906 scope.go:117] "RemoveContainer" containerID="dae906aff5a2080ffb6904c4f1c75020ff73527712718bb4c1180d689e9bae6d" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.090770 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.094754 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.096959 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.100646 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.100886 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.101313 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.101519 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.110481 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.132828 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.145265 4906 scope.go:117] "RemoveContainer" containerID="a7efd54a95aafa3d3cc917715d3b8d6474962d24f6d6688799722b7635d6e399" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.168956 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.169003 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.169059 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckv26\" (UniqueName: \"kubernetes.io/projected/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-kube-api-access-ckv26\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.169129 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.169209 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.169246 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.169347 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.169379 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.270578 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271043 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271077 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-log-httpd\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271141 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271165 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271213 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-run-httpd\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271234 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhh2l\" (UniqueName: \"kubernetes.io/projected/4482bec2-dac8-474e-ba91-12a719d43117-kube-api-access-zhh2l\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271641 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271661 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-config-data\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271724 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-scripts\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271768 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271800 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271876 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckv26\" (UniqueName: \"kubernetes.io/projected/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-kube-api-access-ckv26\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271890 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.271908 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.272012 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.273131 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.273265 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.292389 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.292451 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.293480 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.294060 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.294335 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckv26\" (UniqueName: \"kubernetes.io/projected/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-kube-api-access-ckv26\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.304966 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.373929 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-run-httpd\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.373975 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhh2l\" (UniqueName: \"kubernetes.io/projected/4482bec2-dac8-474e-ba91-12a719d43117-kube-api-access-zhh2l\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.374016 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-config-data\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.374035 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-scripts\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.374078 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.374118 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.374157 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-log-httpd\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.374816 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-log-httpd\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.376489 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-run-httpd\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.379383 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-config-data\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.379700 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-scripts\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.383815 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.390053 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhh2l\" (UniqueName: \"kubernetes.io/projected/4482bec2-dac8-474e-ba91-12a719d43117-kube-api-access-zhh2l\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.390098 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.394890 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.441241 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.474864 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-internal-tls-certs\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.474989 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-httpd-run\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.475034 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-combined-ca-bundle\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.475090 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.475173 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-config-data\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.475229 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcmpg\" (UniqueName: \"kubernetes.io/projected/9e594b90-2851-4374-8913-08103bb5065e-kube-api-access-zcmpg\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.475271 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-logs\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.475290 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-scripts\") pod \"9e594b90-2851-4374-8913-08103bb5065e\" (UID: \"9e594b90-2851-4374-8913-08103bb5065e\") " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.476136 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.476465 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-logs" (OuterVolumeSpecName: "logs") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.478578 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.482140 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e594b90-2851-4374-8913-08103bb5065e-kube-api-access-zcmpg" (OuterVolumeSpecName: "kube-api-access-zcmpg") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "kube-api-access-zcmpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.482666 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-scripts" (OuterVolumeSpecName: "scripts") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.510471 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.549540 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.551382 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-config-data" (OuterVolumeSpecName: "config-data") pod "9e594b90-2851-4374-8913-08103bb5065e" (UID: "9e594b90-2851-4374-8913-08103bb5065e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577130 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577167 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcmpg\" (UniqueName: \"kubernetes.io/projected/9e594b90-2851-4374-8913-08103bb5065e-kube-api-access-zcmpg\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577179 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577189 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577198 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577207 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9e594b90-2851-4374-8913-08103bb5065e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577216 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e594b90-2851-4374-8913-08103bb5065e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.577254 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.598744 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.679265 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.686735 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.961418 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9e594b90-2851-4374-8913-08103bb5065e","Type":"ContainerDied","Data":"883e3792178ea678c841a89a6558d5631e7685262cc51a58ddfe8c25d79faad4"} Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.961930 4906 scope.go:117] "RemoveContainer" containerID="20b05ca7f58640a43e4cde5f3c87862d66748a5431dc1655c260c2a29bf3dae2" Nov 23 07:11:16 crc kubenswrapper[4906]: I1123 07:11:16.962056 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.000902 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.006824 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.015844 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.038307 4906 scope.go:117] "RemoveContainer" containerID="2b9c7e90e2950b3fcc525bdf5a7dbb42733038fec6d65bd737ad2199a9071b46" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.039832 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:11:17 crc kubenswrapper[4906]: E1123 07:11:17.040387 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-log" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.040404 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-log" Nov 23 07:11:17 crc kubenswrapper[4906]: E1123 07:11:17.040626 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-httpd" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.040639 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-httpd" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.041377 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-httpd" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.041444 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e594b90-2851-4374-8913-08103bb5065e" containerName="glance-log" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.043608 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.051610 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.051916 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.061286 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.166357 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189341 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189450 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189478 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189498 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189524 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlvsk\" (UniqueName: \"kubernetes.io/projected/1642db9a-a07d-475a-8112-518d26323d11-kube-api-access-hlvsk\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189574 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-logs\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189615 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.189642 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.290948 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291009 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291032 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291051 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlvsk\" (UniqueName: \"kubernetes.io/projected/1642db9a-a07d-475a-8112-518d26323d11-kube-api-access-hlvsk\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291088 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-logs\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291127 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291153 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291194 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.291846 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.292387 4906 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.292967 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-logs\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.299029 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.299071 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.299430 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.299444 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.312554 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlvsk\" (UniqueName: \"kubernetes.io/projected/1642db9a-a07d-475a-8112-518d26323d11-kube-api-access-hlvsk\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.329622 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.372409 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.394085 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65b17bcd-4689-4bb4-aea9-47288ea27f7d" path="/var/lib/kubelet/pods/65b17bcd-4689-4bb4-aea9-47288ea27f7d/volumes" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.395932 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e594b90-2851-4374-8913-08103bb5065e" path="/var/lib/kubelet/pods/9e594b90-2851-4374-8913-08103bb5065e/volumes" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.397277 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6209532-4ce6-482f-9c69-021b4ec0b682" path="/var/lib/kubelet/pods/e6209532-4ce6-482f-9c69-021b4ec0b682/volumes" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.624787 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2jcp"] Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.626476 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.629327 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nv7gv" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.631395 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.631436 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.640874 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2jcp"] Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.705716 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-scripts\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.705849 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-config-data\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.705892 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.705914 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gl8t\" (UniqueName: \"kubernetes.io/projected/7be07491-e9f1-4b75-9261-c3f49dacb542-kube-api-access-2gl8t\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.809743 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-scripts\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.809880 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-config-data\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.809925 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.809945 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gl8t\" (UniqueName: \"kubernetes.io/projected/7be07491-e9f1-4b75-9261-c3f49dacb542-kube-api-access-2gl8t\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.847226 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.855263 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gl8t\" (UniqueName: \"kubernetes.io/projected/7be07491-e9f1-4b75-9261-c3f49dacb542-kube-api-access-2gl8t\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.855911 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-config-data\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.872314 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-scripts\") pod \"nova-cell0-conductor-db-sync-s2jcp\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.961154 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.978830 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerStarted","Data":"805cca07f3777aaf4b8584d2b2913d1424321480ee38ab697fbcea30233920eb"} Nov 23 07:11:17 crc kubenswrapper[4906]: I1123 07:11:17.981120 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d","Type":"ContainerStarted","Data":"da03f1fec6165233dbdd91e71071a0ec9a3f6fb4ff76060cfa558c2ecdcd499d"} Nov 23 07:11:18 crc kubenswrapper[4906]: I1123 07:11:18.091235 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:11:18 crc kubenswrapper[4906]: I1123 07:11:18.522641 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2jcp"] Nov 23 07:11:18 crc kubenswrapper[4906]: W1123 07:11:18.527060 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7be07491_e9f1_4b75_9261_c3f49dacb542.slice/crio-0ca09c3bbbc2909fa247758dd3d96bddf14468226d66f5bffcaadd483db5f255 WatchSource:0}: Error finding container 0ca09c3bbbc2909fa247758dd3d96bddf14468226d66f5bffcaadd483db5f255: Status 404 returned error can't find the container with id 0ca09c3bbbc2909fa247758dd3d96bddf14468226d66f5bffcaadd483db5f255 Nov 23 07:11:18 crc kubenswrapper[4906]: I1123 07:11:18.995540 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerStarted","Data":"aa8126291bff144f603418e592b1f6dcaccc4c065fbe8568ca377e65087939b3"} Nov 23 07:11:18 crc kubenswrapper[4906]: I1123 07:11:18.997783 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d","Type":"ContainerStarted","Data":"9c4e48b39c14772f337cfc58ea53a79082fd8c022fe4b10170edefdfb0a64859"} Nov 23 07:11:18 crc kubenswrapper[4906]: I1123 07:11:18.999439 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" event={"ID":"7be07491-e9f1-4b75-9261-c3f49dacb542","Type":"ContainerStarted","Data":"0ca09c3bbbc2909fa247758dd3d96bddf14468226d66f5bffcaadd483db5f255"} Nov 23 07:11:19 crc kubenswrapper[4906]: I1123 07:11:19.003442 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1642db9a-a07d-475a-8112-518d26323d11","Type":"ContainerStarted","Data":"116a985e4f0a6ad1024c63bfe365d28dafd484ecb5fc5da136db31844fa21b2b"} Nov 23 07:11:22 crc kubenswrapper[4906]: I1123 07:11:22.036541 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1642db9a-a07d-475a-8112-518d26323d11","Type":"ContainerStarted","Data":"f4c66a184ddf3fddd4f01f4bf5d9927e08a50644685ee93df83ca8a705617d12"} Nov 23 07:11:22 crc kubenswrapper[4906]: I1123 07:11:22.037521 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1642db9a-a07d-475a-8112-518d26323d11","Type":"ContainerStarted","Data":"85604e28eb73da42df364a73f2a404a0c2cac5adb3cdbd3da52bb12fba8bccc0"} Nov 23 07:11:22 crc kubenswrapper[4906]: I1123 07:11:22.043996 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerStarted","Data":"2c8ea7bf763beb872aa76646d1aa32281a56e75ac152c2352a663e16c3a97454"} Nov 23 07:11:22 crc kubenswrapper[4906]: I1123 07:11:22.048847 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d","Type":"ContainerStarted","Data":"2fd16719fcd506b038a0152208792ed142945112e8297ebc99ad99b173d43f11"} Nov 23 07:11:22 crc kubenswrapper[4906]: I1123 07:11:22.068633 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.068600652 podStartE2EDuration="5.068600652s" podCreationTimestamp="2025-11-23 07:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:11:22.057715166 +0000 UTC m=+1297.571106499" watchObservedRunningTime="2025-11-23 07:11:22.068600652 +0000 UTC m=+1297.581991965" Nov 23 07:11:23 crc kubenswrapper[4906]: I1123 07:11:23.065264 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerStarted","Data":"c76a5098fb2971db8ea7b34b60e4319599ca18452356b9c29903821f2d98dc07"} Nov 23 07:11:25 crc kubenswrapper[4906]: I1123 07:11:25.396324 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.396296516 podStartE2EDuration="10.396296516s" podCreationTimestamp="2025-11-23 07:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:11:22.081835298 +0000 UTC m=+1297.595226611" watchObservedRunningTime="2025-11-23 07:11:25.396296516 +0000 UTC m=+1300.909687819" Nov 23 07:11:26 crc kubenswrapper[4906]: I1123 07:11:26.442203 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 07:11:26 crc kubenswrapper[4906]: I1123 07:11:26.442726 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 07:11:26 crc kubenswrapper[4906]: I1123 07:11:26.475714 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 07:11:26 crc kubenswrapper[4906]: I1123 07:11:26.485103 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 07:11:27 crc kubenswrapper[4906]: I1123 07:11:27.113758 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 07:11:27 crc kubenswrapper[4906]: I1123 07:11:27.114364 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 07:11:27 crc kubenswrapper[4906]: I1123 07:11:27.376589 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:27 crc kubenswrapper[4906]: I1123 07:11:27.376656 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:27 crc kubenswrapper[4906]: I1123 07:11:27.426029 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:27 crc kubenswrapper[4906]: I1123 07:11:27.477501 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:28 crc kubenswrapper[4906]: I1123 07:11:28.130544 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:28 crc kubenswrapper[4906]: I1123 07:11:28.130957 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:29 crc kubenswrapper[4906]: I1123 07:11:29.153395 4906 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 07:11:29 crc kubenswrapper[4906]: I1123 07:11:29.153835 4906 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 07:11:29 crc kubenswrapper[4906]: I1123 07:11:29.153399 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerStarted","Data":"e67f16d70428889059a4cd8f993f06a74b72e1768492d989fa9a429a6fc5133b"} Nov 23 07:11:29 crc kubenswrapper[4906]: I1123 07:11:29.153947 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:11:29 crc kubenswrapper[4906]: I1123 07:11:29.154406 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 07:11:29 crc kubenswrapper[4906]: I1123 07:11:29.154842 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 07:11:29 crc kubenswrapper[4906]: I1123 07:11:29.187119 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.962829173 podStartE2EDuration="13.187096156s" podCreationTimestamp="2025-11-23 07:11:16 +0000 UTC" firstStartedPulling="2025-11-23 07:11:17.175288993 +0000 UTC m=+1292.688680296" lastFinishedPulling="2025-11-23 07:11:23.399555966 +0000 UTC m=+1298.912947279" observedRunningTime="2025-11-23 07:11:29.17392608 +0000 UTC m=+1304.687317393" watchObservedRunningTime="2025-11-23 07:11:29.187096156 +0000 UTC m=+1304.700487459" Nov 23 07:11:30 crc kubenswrapper[4906]: I1123 07:11:30.112261 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:30 crc kubenswrapper[4906]: I1123 07:11:30.114880 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 07:11:31 crc kubenswrapper[4906]: I1123 07:11:31.173700 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" event={"ID":"7be07491-e9f1-4b75-9261-c3f49dacb542","Type":"ContainerStarted","Data":"a1ce0f6ba41a9ea922831c56c6043771d63ea9f48abecf10c7758bee8f51ec7e"} Nov 23 07:11:31 crc kubenswrapper[4906]: I1123 07:11:31.197802 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" podStartSLOduration=2.748179699 podStartE2EDuration="14.197777021s" podCreationTimestamp="2025-11-23 07:11:17 +0000 UTC" firstStartedPulling="2025-11-23 07:11:18.530133293 +0000 UTC m=+1294.043524596" lastFinishedPulling="2025-11-23 07:11:29.979730595 +0000 UTC m=+1305.493121918" observedRunningTime="2025-11-23 07:11:31.191560597 +0000 UTC m=+1306.704951900" watchObservedRunningTime="2025-11-23 07:11:31.197777021 +0000 UTC m=+1306.711168324" Nov 23 07:11:44 crc kubenswrapper[4906]: I1123 07:11:44.347584 4906 generic.go:334] "Generic (PLEG): container finished" podID="7be07491-e9f1-4b75-9261-c3f49dacb542" containerID="a1ce0f6ba41a9ea922831c56c6043771d63ea9f48abecf10c7758bee8f51ec7e" exitCode=0 Nov 23 07:11:44 crc kubenswrapper[4906]: I1123 07:11:44.347656 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" event={"ID":"7be07491-e9f1-4b75-9261-c3f49dacb542","Type":"ContainerDied","Data":"a1ce0f6ba41a9ea922831c56c6043771d63ea9f48abecf10c7758bee8f51ec7e"} Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.827303 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.945148 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-combined-ca-bundle\") pod \"7be07491-e9f1-4b75-9261-c3f49dacb542\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.945224 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gl8t\" (UniqueName: \"kubernetes.io/projected/7be07491-e9f1-4b75-9261-c3f49dacb542-kube-api-access-2gl8t\") pod \"7be07491-e9f1-4b75-9261-c3f49dacb542\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.945333 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-config-data\") pod \"7be07491-e9f1-4b75-9261-c3f49dacb542\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.945487 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-scripts\") pod \"7be07491-e9f1-4b75-9261-c3f49dacb542\" (UID: \"7be07491-e9f1-4b75-9261-c3f49dacb542\") " Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.956964 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7be07491-e9f1-4b75-9261-c3f49dacb542-kube-api-access-2gl8t" (OuterVolumeSpecName: "kube-api-access-2gl8t") pod "7be07491-e9f1-4b75-9261-c3f49dacb542" (UID: "7be07491-e9f1-4b75-9261-c3f49dacb542"). InnerVolumeSpecName "kube-api-access-2gl8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.959593 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-scripts" (OuterVolumeSpecName: "scripts") pod "7be07491-e9f1-4b75-9261-c3f49dacb542" (UID: "7be07491-e9f1-4b75-9261-c3f49dacb542"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.973541 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-config-data" (OuterVolumeSpecName: "config-data") pod "7be07491-e9f1-4b75-9261-c3f49dacb542" (UID: "7be07491-e9f1-4b75-9261-c3f49dacb542"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:45 crc kubenswrapper[4906]: I1123 07:11:45.975197 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7be07491-e9f1-4b75-9261-c3f49dacb542" (UID: "7be07491-e9f1-4b75-9261-c3f49dacb542"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.048186 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.048230 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gl8t\" (UniqueName: \"kubernetes.io/projected/7be07491-e9f1-4b75-9261-c3f49dacb542-kube-api-access-2gl8t\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.048246 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.048261 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7be07491-e9f1-4b75-9261-c3f49dacb542-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.377481 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" event={"ID":"7be07491-e9f1-4b75-9261-c3f49dacb542","Type":"ContainerDied","Data":"0ca09c3bbbc2909fa247758dd3d96bddf14468226d66f5bffcaadd483db5f255"} Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.377993 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ca09c3bbbc2909fa247758dd3d96bddf14468226d66f5bffcaadd483db5f255" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.377613 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s2jcp" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.545885 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:11:46 crc kubenswrapper[4906]: E1123 07:11:46.546549 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be07491-e9f1-4b75-9261-c3f49dacb542" containerName="nova-cell0-conductor-db-sync" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.546578 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be07491-e9f1-4b75-9261-c3f49dacb542" containerName="nova-cell0-conductor-db-sync" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.546868 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be07491-e9f1-4b75-9261-c3f49dacb542" containerName="nova-cell0-conductor-db-sync" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.547875 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.551554 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.555527 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nv7gv" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.563203 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.665417 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.666096 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfh95\" (UniqueName: \"kubernetes.io/projected/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-kube-api-access-lfh95\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.666369 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.694264 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.769121 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.769942 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfh95\" (UniqueName: \"kubernetes.io/projected/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-kube-api-access-lfh95\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.770092 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.780348 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.784510 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.795821 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfh95\" (UniqueName: \"kubernetes.io/projected/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-kube-api-access-lfh95\") pod \"nova-cell0-conductor-0\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:46 crc kubenswrapper[4906]: I1123 07:11:46.867588 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:47 crc kubenswrapper[4906]: I1123 07:11:47.374392 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:11:47 crc kubenswrapper[4906]: W1123 07:11:47.377395 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ab3c4dd_5b70_4127_aa8d_570ffcde7477.slice/crio-9649c64a7d519f1db26ea6b55d1a041e210285132102870eff86164c7ebc9e8a WatchSource:0}: Error finding container 9649c64a7d519f1db26ea6b55d1a041e210285132102870eff86164c7ebc9e8a: Status 404 returned error can't find the container with id 9649c64a7d519f1db26ea6b55d1a041e210285132102870eff86164c7ebc9e8a Nov 23 07:11:48 crc kubenswrapper[4906]: I1123 07:11:48.412074 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1ab3c4dd-5b70-4127-aa8d-570ffcde7477","Type":"ContainerStarted","Data":"a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33"} Nov 23 07:11:48 crc kubenswrapper[4906]: I1123 07:11:48.412662 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1ab3c4dd-5b70-4127-aa8d-570ffcde7477","Type":"ContainerStarted","Data":"9649c64a7d519f1db26ea6b55d1a041e210285132102870eff86164c7ebc9e8a"} Nov 23 07:11:48 crc kubenswrapper[4906]: I1123 07:11:48.412690 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:48 crc kubenswrapper[4906]: I1123 07:11:48.438765 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.436903941 podStartE2EDuration="2.436903941s" podCreationTimestamp="2025-11-23 07:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:11:48.428738207 +0000 UTC m=+1323.942129510" watchObservedRunningTime="2025-11-23 07:11:48.436903941 +0000 UTC m=+1323.950295244" Nov 23 07:11:50 crc kubenswrapper[4906]: I1123 07:11:50.841203 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:11:50 crc kubenswrapper[4906]: I1123 07:11:50.841923 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" containerName="kube-state-metrics" containerID="cri-o://3d34a0cfbf6cce02048082a86c433bac46bdbc467dfe235995d608ebba0146d0" gracePeriod=30 Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.447162 4906 generic.go:334] "Generic (PLEG): container finished" podID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" containerID="3d34a0cfbf6cce02048082a86c433bac46bdbc467dfe235995d608ebba0146d0" exitCode=2 Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.447234 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e368b3d0-8ac9-49be-b2c4-8e88d8327811","Type":"ContainerDied","Data":"3d34a0cfbf6cce02048082a86c433bac46bdbc467dfe235995d608ebba0146d0"} Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.447589 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e368b3d0-8ac9-49be-b2c4-8e88d8327811","Type":"ContainerDied","Data":"aa2feed880b147f6869bba98d26de9ee9ad67b63a10e56e9017b6a965cd060de"} Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.447618 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa2feed880b147f6869bba98d26de9ee9ad67b63a10e56e9017b6a965cd060de" Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.464828 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.574384 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f25fh\" (UniqueName: \"kubernetes.io/projected/e368b3d0-8ac9-49be-b2c4-8e88d8327811-kube-api-access-f25fh\") pod \"e368b3d0-8ac9-49be-b2c4-8e88d8327811\" (UID: \"e368b3d0-8ac9-49be-b2c4-8e88d8327811\") " Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.582645 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e368b3d0-8ac9-49be-b2c4-8e88d8327811-kube-api-access-f25fh" (OuterVolumeSpecName: "kube-api-access-f25fh") pod "e368b3d0-8ac9-49be-b2c4-8e88d8327811" (UID: "e368b3d0-8ac9-49be-b2c4-8e88d8327811"). InnerVolumeSpecName "kube-api-access-f25fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:51 crc kubenswrapper[4906]: I1123 07:11:51.678452 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f25fh\" (UniqueName: \"kubernetes.io/projected/e368b3d0-8ac9-49be-b2c4-8e88d8327811-kube-api-access-f25fh\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.455280 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.487096 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.495640 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.513644 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:11:52 crc kubenswrapper[4906]: E1123 07:11:52.514551 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" containerName="kube-state-metrics" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.514587 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" containerName="kube-state-metrics" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.515285 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" containerName="kube-state-metrics" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.516346 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.518183 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.518827 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.525278 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.696963 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6g5l\" (UniqueName: \"kubernetes.io/projected/9535028a-afb6-49d6-9bac-541c5324cc55-kube-api-access-p6g5l\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.697089 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.697218 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.697303 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.702386 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.702720 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-central-agent" containerID="cri-o://aa8126291bff144f603418e592b1f6dcaccc4c065fbe8568ca377e65087939b3" gracePeriod=30 Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.702811 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="sg-core" containerID="cri-o://c76a5098fb2971db8ea7b34b60e4319599ca18452356b9c29903821f2d98dc07" gracePeriod=30 Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.702821 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="proxy-httpd" containerID="cri-o://e67f16d70428889059a4cd8f993f06a74b72e1768492d989fa9a429a6fc5133b" gracePeriod=30 Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.702847 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-notification-agent" containerID="cri-o://2c8ea7bf763beb872aa76646d1aa32281a56e75ac152c2352a663e16c3a97454" gracePeriod=30 Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.798835 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.798947 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.799102 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6g5l\" (UniqueName: \"kubernetes.io/projected/9535028a-afb6-49d6-9bac-541c5324cc55-kube-api-access-p6g5l\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.799158 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.804516 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.807274 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.811309 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.819365 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6g5l\" (UniqueName: \"kubernetes.io/projected/9535028a-afb6-49d6-9bac-541c5324cc55-kube-api-access-p6g5l\") pod \"kube-state-metrics-0\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " pod="openstack/kube-state-metrics-0" Nov 23 07:11:52 crc kubenswrapper[4906]: I1123 07:11:52.839814 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.372696 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e368b3d0-8ac9-49be-b2c4-8e88d8327811" path="/var/lib/kubelet/pods/e368b3d0-8ac9-49be-b2c4-8e88d8327811/volumes" Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.373913 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.471279 4906 generic.go:334] "Generic (PLEG): container finished" podID="4482bec2-dac8-474e-ba91-12a719d43117" containerID="e67f16d70428889059a4cd8f993f06a74b72e1768492d989fa9a429a6fc5133b" exitCode=0 Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.472324 4906 generic.go:334] "Generic (PLEG): container finished" podID="4482bec2-dac8-474e-ba91-12a719d43117" containerID="c76a5098fb2971db8ea7b34b60e4319599ca18452356b9c29903821f2d98dc07" exitCode=2 Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.472429 4906 generic.go:334] "Generic (PLEG): container finished" podID="4482bec2-dac8-474e-ba91-12a719d43117" containerID="aa8126291bff144f603418e592b1f6dcaccc4c065fbe8568ca377e65087939b3" exitCode=0 Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.471371 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerDied","Data":"e67f16d70428889059a4cd8f993f06a74b72e1768492d989fa9a429a6fc5133b"} Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.472648 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerDied","Data":"c76a5098fb2971db8ea7b34b60e4319599ca18452356b9c29903821f2d98dc07"} Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.472668 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerDied","Data":"aa8126291bff144f603418e592b1f6dcaccc4c065fbe8568ca377e65087939b3"} Nov 23 07:11:53 crc kubenswrapper[4906]: I1123 07:11:53.474921 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9535028a-afb6-49d6-9bac-541c5324cc55","Type":"ContainerStarted","Data":"11579df96862d0b66237165e68aa09a4c23bfea76d77ad2076e3e62b0c745db8"} Nov 23 07:11:54 crc kubenswrapper[4906]: I1123 07:11:54.488368 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9535028a-afb6-49d6-9bac-541c5324cc55","Type":"ContainerStarted","Data":"0ef240d12f0659c3ea6a9b32c942f26f9370dad6938a45b1af45fb31d908312f"} Nov 23 07:11:54 crc kubenswrapper[4906]: I1123 07:11:54.488927 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 23 07:11:54 crc kubenswrapper[4906]: I1123 07:11:54.516938 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.160957976 podStartE2EDuration="2.516904904s" podCreationTimestamp="2025-11-23 07:11:52 +0000 UTC" firstStartedPulling="2025-11-23 07:11:53.375428753 +0000 UTC m=+1328.888820056" lastFinishedPulling="2025-11-23 07:11:53.731375681 +0000 UTC m=+1329.244766984" observedRunningTime="2025-11-23 07:11:54.510114666 +0000 UTC m=+1330.023506019" watchObservedRunningTime="2025-11-23 07:11:54.516904904 +0000 UTC m=+1330.030296237" Nov 23 07:11:56 crc kubenswrapper[4906]: I1123 07:11:56.903854 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.428474 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-h4lvc"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.437139 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.439885 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.440497 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.454377 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-h4lvc"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.542595 4906 generic.go:334] "Generic (PLEG): container finished" podID="4482bec2-dac8-474e-ba91-12a719d43117" containerID="2c8ea7bf763beb872aa76646d1aa32281a56e75ac152c2352a663e16c3a97454" exitCode=0 Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.542645 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerDied","Data":"2c8ea7bf763beb872aa76646d1aa32281a56e75ac152c2352a663e16c3a97454"} Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.542710 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4482bec2-dac8-474e-ba91-12a719d43117","Type":"ContainerDied","Data":"805cca07f3777aaf4b8584d2b2913d1424321480ee38ab697fbcea30233920eb"} Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.542727 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="805cca07f3777aaf4b8584d2b2913d1424321480ee38ab697fbcea30233920eb" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.588929 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.589476 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:11:57 crc kubenswrapper[4906]: E1123 07:11:57.591908 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="proxy-httpd" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.591937 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="proxy-httpd" Nov 23 07:11:57 crc kubenswrapper[4906]: E1123 07:11:57.591975 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-central-agent" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.591982 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-central-agent" Nov 23 07:11:57 crc kubenswrapper[4906]: E1123 07:11:57.592032 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="sg-core" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.592041 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="sg-core" Nov 23 07:11:57 crc kubenswrapper[4906]: E1123 07:11:57.594090 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-notification-agent" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.594127 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-notification-agent" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.594352 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-notification-agent" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.594377 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="ceilometer-central-agent" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.594387 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="proxy-httpd" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.594402 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4482bec2-dac8-474e-ba91-12a719d43117" containerName="sg-core" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.595232 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.602711 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628383 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-combined-ca-bundle\") pod \"4482bec2-dac8-474e-ba91-12a719d43117\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628453 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-sg-core-conf-yaml\") pod \"4482bec2-dac8-474e-ba91-12a719d43117\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628510 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-config-data\") pod \"4482bec2-dac8-474e-ba91-12a719d43117\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628549 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-run-httpd\") pod \"4482bec2-dac8-474e-ba91-12a719d43117\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628601 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-scripts\") pod \"4482bec2-dac8-474e-ba91-12a719d43117\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628642 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-log-httpd\") pod \"4482bec2-dac8-474e-ba91-12a719d43117\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628715 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhh2l\" (UniqueName: \"kubernetes.io/projected/4482bec2-dac8-474e-ba91-12a719d43117-kube-api-access-zhh2l\") pod \"4482bec2-dac8-474e-ba91-12a719d43117\" (UID: \"4482bec2-dac8-474e-ba91-12a719d43117\") " Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628855 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kf44\" (UniqueName: \"kubernetes.io/projected/35e15932-ac43-4d8e-ba00-481ad71b5816-kube-api-access-6kf44\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628922 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnr8n\" (UniqueName: \"kubernetes.io/projected/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-kube-api-access-lnr8n\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.628977 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.629130 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-scripts\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.629215 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-config-data\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.629245 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-config-data\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.629308 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.630744 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4482bec2-dac8-474e-ba91-12a719d43117" (UID: "4482bec2-dac8-474e-ba91-12a719d43117"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.634891 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4482bec2-dac8-474e-ba91-12a719d43117" (UID: "4482bec2-dac8-474e-ba91-12a719d43117"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.642378 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.653034 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4482bec2-dac8-474e-ba91-12a719d43117-kube-api-access-zhh2l" (OuterVolumeSpecName: "kube-api-access-zhh2l") pod "4482bec2-dac8-474e-ba91-12a719d43117" (UID: "4482bec2-dac8-474e-ba91-12a719d43117"). InnerVolumeSpecName "kube-api-access-zhh2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.661080 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-scripts" (OuterVolumeSpecName: "scripts") pod "4482bec2-dac8-474e-ba91-12a719d43117" (UID: "4482bec2-dac8-474e-ba91-12a719d43117"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.719513 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.726140 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733150 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-config-data\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733227 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733250 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kf44\" (UniqueName: \"kubernetes.io/projected/35e15932-ac43-4d8e-ba00-481ad71b5816-kube-api-access-6kf44\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733290 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q692w\" (UniqueName: \"kubernetes.io/projected/10d92e7e-6c63-41b9-83d7-663163ee7a00-kube-api-access-q692w\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733310 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnr8n\" (UniqueName: \"kubernetes.io/projected/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-kube-api-access-lnr8n\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733328 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733357 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733386 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-scripts\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733402 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10d92e7e-6c63-41b9-83d7-663163ee7a00-logs\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733458 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-config-data\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733486 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-config-data\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733576 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733590 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733600 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4482bec2-dac8-474e-ba91-12a719d43117-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.733613 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhh2l\" (UniqueName: \"kubernetes.io/projected/4482bec2-dac8-474e-ba91-12a719d43117-kube-api-access-zhh2l\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.734790 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.765025 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnr8n\" (UniqueName: \"kubernetes.io/projected/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-kube-api-access-lnr8n\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.769491 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-config-data\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.769823 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4482bec2-dac8-474e-ba91-12a719d43117" (UID: "4482bec2-dac8-474e-ba91-12a719d43117"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.770747 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.772799 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-scripts\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.779073 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.784095 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-config-data\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.784756 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kf44\" (UniqueName: \"kubernetes.io/projected/35e15932-ac43-4d8e-ba00-481ad71b5816-kube-api-access-6kf44\") pod \"nova-cell0-cell-mapping-h4lvc\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.816352 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.822512 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.841876 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q692w\" (UniqueName: \"kubernetes.io/projected/10d92e7e-6c63-41b9-83d7-663163ee7a00-kube-api-access-q692w\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.841951 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.841992 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10d92e7e-6c63-41b9-83d7-663163ee7a00-logs\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.842081 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-config-data\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.871393 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.881018 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4482bec2-dac8-474e-ba91-12a719d43117" (UID: "4482bec2-dac8-474e-ba91-12a719d43117"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.884890 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.903944 4906 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.908302 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.912659 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.913338 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10d92e7e-6c63-41b9-83d7-663163ee7a00-logs\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.931617 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.955334 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-config-data\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.961894 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q692w\" (UniqueName: \"kubernetes.io/projected/10d92e7e-6c63-41b9-83d7-663163ee7a00-kube-api-access-q692w\") pod \"nova-metadata-0\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " pod="openstack/nova-metadata-0" Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.984163 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.993216 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-lwpwn"] Nov 23 07:11:57 crc kubenswrapper[4906]: I1123 07:11:57.995291 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.003027 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.005824 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fmpk\" (UniqueName: \"kubernetes.io/projected/075c468b-da1c-4e30-b4cf-fe52cb94a76a-kube-api-access-5fmpk\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.005875 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-config-data\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.005942 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075c468b-da1c-4e30-b4cf-fe52cb94a76a-logs\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.005963 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.006869 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.008391 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.012032 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.013500 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-config-data" (OuterVolumeSpecName: "config-data") pod "4482bec2-dac8-474e-ba91-12a719d43117" (UID: "4482bec2-dac8-474e-ba91-12a719d43117"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.025742 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-lwpwn"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.043949 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.087410 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.110571 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111077 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111138 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjrs9\" (UniqueName: \"kubernetes.io/projected/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-kube-api-access-mjrs9\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111163 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111222 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fmpk\" (UniqueName: \"kubernetes.io/projected/075c468b-da1c-4e30-b4cf-fe52cb94a76a-kube-api-access-5fmpk\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111248 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-config-data\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111266 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g5hx\" (UniqueName: \"kubernetes.io/projected/ad9b24a7-201a-4c2d-8082-8b911f70b76e-kube-api-access-2g5hx\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111320 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111386 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075c468b-da1c-4e30-b4cf-fe52cb94a76a-logs\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111408 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111449 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111486 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111503 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-config\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.111582 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4482bec2-dac8-474e-ba91-12a719d43117-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.113536 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075c468b-da1c-4e30-b4cf-fe52cb94a76a-logs\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.119783 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.145900 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fmpk\" (UniqueName: \"kubernetes.io/projected/075c468b-da1c-4e30-b4cf-fe52cb94a76a-kube-api-access-5fmpk\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.146047 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-config-data\") pod \"nova-api-0\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214365 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjrs9\" (UniqueName: \"kubernetes.io/projected/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-kube-api-access-mjrs9\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214440 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214484 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g5hx\" (UniqueName: \"kubernetes.io/projected/ad9b24a7-201a-4c2d-8082-8b911f70b76e-kube-api-access-2g5hx\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214521 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214569 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214608 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214629 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-config\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214666 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.214734 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.215940 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.216025 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-config\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.216431 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.216609 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.221398 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.243739 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.245178 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.249667 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g5hx\" (UniqueName: \"kubernetes.io/projected/ad9b24a7-201a-4c2d-8082-8b911f70b76e-kube-api-access-2g5hx\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.253283 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjrs9\" (UniqueName: \"kubernetes.io/projected/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-kube-api-access-mjrs9\") pod \"dnsmasq-dns-5dd7c4987f-lwpwn\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.255335 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.344735 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-h4lvc"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.357414 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.361565 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.558022 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.562255 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h4lvc" event={"ID":"35e15932-ac43-4d8e-ba00-481ad71b5816","Type":"ContainerStarted","Data":"881106b2832b86c7da98f8a1a7b4b1b987139699039512e4a1540be32cb511e1"} Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.604903 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.620273 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.632072 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.673835 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.673961 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.676147 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:11:58 crc kubenswrapper[4906]: E1123 07:11:58.676948 4906 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4482bec2_dac8_474e_ba91_12a719d43117.slice\": RecentStats: unable to find data in memory cache]" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.677189 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.677380 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.693421 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.801314 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: W1123 07:11:58.821447 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10d92e7e_6c63_41b9_83d7_663163ee7a00.slice/crio-53df4b146e1e01f787ac1466a20848ad875bdf3599de03714bf4557db8547212 WatchSource:0}: Error finding container 53df4b146e1e01f787ac1466a20848ad875bdf3599de03714bf4557db8547212: Status 404 returned error can't find the container with id 53df4b146e1e01f787ac1466a20848ad875bdf3599de03714bf4557db8547212 Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.830234 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.830868 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4sw6\" (UniqueName: \"kubernetes.io/projected/a91aaae7-e5f2-4f11-b3e4-a15179305476-kube-api-access-j4sw6\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.830943 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.831018 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-log-httpd\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.831065 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.831189 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-scripts\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.831263 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.831302 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-config-data\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.831342 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-run-httpd\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933017 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4sw6\" (UniqueName: \"kubernetes.io/projected/a91aaae7-e5f2-4f11-b3e4-a15179305476-kube-api-access-j4sw6\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933495 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933559 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-log-httpd\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933614 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933712 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-scripts\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933751 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933791 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-config-data\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.933806 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-run-httpd\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.934333 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-run-httpd\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.935554 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-log-httpd\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.941278 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.941593 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-scripts\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.943475 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.944307 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.948610 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-config-data\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.955543 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4sw6\" (UniqueName: \"kubernetes.io/projected/a91aaae7-e5f2-4f11-b3e4-a15179305476-kube-api-access-j4sw6\") pod \"ceilometer-0\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " pod="openstack/ceilometer-0" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.985336 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-54gvc"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.986893 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.991174 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-54gvc"] Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.991322 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 23 07:11:58 crc kubenswrapper[4906]: I1123 07:11:58.993059 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.004442 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.020171 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:11:59 crc kubenswrapper[4906]: W1123 07:11:59.106399 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod009ce17b_acb9_47b7_b4ed_6ebca0ef635a.slice/crio-128a8304c71f69c84f55fbdf15e6046ac6779a39aaf7b2808a4fd17d9d694a33 WatchSource:0}: Error finding container 128a8304c71f69c84f55fbdf15e6046ac6779a39aaf7b2808a4fd17d9d694a33: Status 404 returned error can't find the container with id 128a8304c71f69c84f55fbdf15e6046ac6779a39aaf7b2808a4fd17d9d694a33 Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.109079 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-lwpwn"] Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.138639 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-config-data\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.138728 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.138804 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-scripts\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.138836 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gclt9\" (UniqueName: \"kubernetes.io/projected/df30375c-b83c-490c-be3d-2453779a3c06-kube-api-access-gclt9\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.241787 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-config-data\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.242144 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.242208 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-scripts\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.242247 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gclt9\" (UniqueName: \"kubernetes.io/projected/df30375c-b83c-490c-be3d-2453779a3c06-kube-api-access-gclt9\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.249473 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-scripts\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.249792 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-config-data\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.250887 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.264325 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gclt9\" (UniqueName: \"kubernetes.io/projected/df30375c-b83c-490c-be3d-2453779a3c06-kube-api-access-gclt9\") pod \"nova-cell1-conductor-db-sync-54gvc\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.310837 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.385566 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4482bec2-dac8-474e-ba91-12a719d43117" path="/var/lib/kubelet/pods/4482bec2-dac8-474e-ba91-12a719d43117/volumes" Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.584240 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.593104 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"075c468b-da1c-4e30-b4cf-fe52cb94a76a","Type":"ContainerStarted","Data":"62ed8be502476a60d6da1cd8eb02485b292dbcbe09ad30956472bde7e69c2894"} Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.594872 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1de8803b-196b-4eae-8b9e-e9dd02e12a7d","Type":"ContainerStarted","Data":"dacfc102737d204884a1a66c6a0bb44015d08fcb5123629fbfde147482ae46c5"} Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.604522 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10d92e7e-6c63-41b9-83d7-663163ee7a00","Type":"ContainerStarted","Data":"53df4b146e1e01f787ac1466a20848ad875bdf3599de03714bf4557db8547212"} Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.620057 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad9b24a7-201a-4c2d-8082-8b911f70b76e","Type":"ContainerStarted","Data":"e1ebce5afce804edbb82b4b6a67f7646dfd885bbdf590363528b40fae34ff0fe"} Nov 23 07:11:59 crc kubenswrapper[4906]: W1123 07:11:59.629736 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda91aaae7_e5f2_4f11_b3e4_a15179305476.slice/crio-1b35615d4f7e05c5cfb605d405a85cc02c37f8ad0235c0aab47be954cbf06368 WatchSource:0}: Error finding container 1b35615d4f7e05c5cfb605d405a85cc02c37f8ad0235c0aab47be954cbf06368: Status 404 returned error can't find the container with id 1b35615d4f7e05c5cfb605d405a85cc02c37f8ad0235c0aab47be954cbf06368 Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.635082 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" event={"ID":"009ce17b-acb9-47b7-b4ed-6ebca0ef635a","Type":"ContainerStarted","Data":"aee41b6a5dd1ef9d39f6a37ecd59b25e04cf324a7f6135066b76c4a4552915f9"} Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.635166 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" event={"ID":"009ce17b-acb9-47b7-b4ed-6ebca0ef635a","Type":"ContainerStarted","Data":"128a8304c71f69c84f55fbdf15e6046ac6779a39aaf7b2808a4fd17d9d694a33"} Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.661775 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h4lvc" event={"ID":"35e15932-ac43-4d8e-ba00-481ad71b5816","Type":"ContainerStarted","Data":"6f1bb39f61ceebf4bbb9cb405575f82b1dff5097833cdf28d70e297480072af0"} Nov 23 07:11:59 crc kubenswrapper[4906]: I1123 07:11:59.705140 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-h4lvc" podStartSLOduration=2.705116449 podStartE2EDuration="2.705116449s" podCreationTimestamp="2025-11-23 07:11:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:11:59.692280643 +0000 UTC m=+1335.205671946" watchObservedRunningTime="2025-11-23 07:11:59.705116449 +0000 UTC m=+1335.218507752" Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.001524 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-54gvc"] Nov 23 07:12:00 crc kubenswrapper[4906]: W1123 07:12:00.012792 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf30375c_b83c_490c_be3d_2453779a3c06.slice/crio-8fa3b71554771f95e8b74960c97318ac51bb50297d161c7478cddb60493f4998 WatchSource:0}: Error finding container 8fa3b71554771f95e8b74960c97318ac51bb50297d161c7478cddb60493f4998: Status 404 returned error can't find the container with id 8fa3b71554771f95e8b74960c97318ac51bb50297d161c7478cddb60493f4998 Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.677269 4906 generic.go:334] "Generic (PLEG): container finished" podID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerID="aee41b6a5dd1ef9d39f6a37ecd59b25e04cf324a7f6135066b76c4a4552915f9" exitCode=0 Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.677363 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" event={"ID":"009ce17b-acb9-47b7-b4ed-6ebca0ef635a","Type":"ContainerDied","Data":"aee41b6a5dd1ef9d39f6a37ecd59b25e04cf324a7f6135066b76c4a4552915f9"} Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.677926 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" event={"ID":"009ce17b-acb9-47b7-b4ed-6ebca0ef635a","Type":"ContainerStarted","Data":"9cf15debcaa784ebb7c366d5efc1dfc07119d5f901f8b3e1db3f4dfeef24c333"} Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.677945 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.681557 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerStarted","Data":"1b35615d4f7e05c5cfb605d405a85cc02c37f8ad0235c0aab47be954cbf06368"} Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.684536 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-54gvc" event={"ID":"df30375c-b83c-490c-be3d-2453779a3c06","Type":"ContainerStarted","Data":"96435a9a6698e54650b138b7b8daa6f2fcc2fc3938e89653a935fea62eeb5cf6"} Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.684596 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-54gvc" event={"ID":"df30375c-b83c-490c-be3d-2453779a3c06","Type":"ContainerStarted","Data":"8fa3b71554771f95e8b74960c97318ac51bb50297d161c7478cddb60493f4998"} Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.705881 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" podStartSLOduration=3.705850711 podStartE2EDuration="3.705850711s" podCreationTimestamp="2025-11-23 07:11:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:00.697734218 +0000 UTC m=+1336.211125531" watchObservedRunningTime="2025-11-23 07:12:00.705850711 +0000 UTC m=+1336.219242014" Nov 23 07:12:00 crc kubenswrapper[4906]: I1123 07:12:00.717437 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-54gvc" podStartSLOduration=2.717412924 podStartE2EDuration="2.717412924s" podCreationTimestamp="2025-11-23 07:11:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:00.713505511 +0000 UTC m=+1336.226896814" watchObservedRunningTime="2025-11-23 07:12:00.717412924 +0000 UTC m=+1336.230804227" Nov 23 07:12:02 crc kubenswrapper[4906]: I1123 07:12:02.022383 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:02 crc kubenswrapper[4906]: I1123 07:12:02.045434 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:12:02 crc kubenswrapper[4906]: I1123 07:12:02.728757 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerStarted","Data":"ce47dd2e8dc264da48a333d312b4f60fdd8a13b1b6573e5fc8dd87fa49c1e4a7"} Nov 23 07:12:02 crc kubenswrapper[4906]: I1123 07:12:02.873189 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.749158 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"075c468b-da1c-4e30-b4cf-fe52cb94a76a","Type":"ContainerStarted","Data":"eea407ed926731b5de4cfa6b5d5d8ba32c637b612478efe45f499b87b1b68fc9"} Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.752627 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"075c468b-da1c-4e30-b4cf-fe52cb94a76a","Type":"ContainerStarted","Data":"2846c1bbb7a7f281e8fd181fee816335cda2a4869820e6a6fccd491b87baf0e7"} Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.761053 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerStarted","Data":"1f4258c9e792706d58c0124470407407625159d5bd411557fb8aa27bf93252e5"} Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.762613 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1de8803b-196b-4eae-8b9e-e9dd02e12a7d","Type":"ContainerStarted","Data":"7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c"} Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.768977 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-log" containerID="cri-o://af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1" gracePeriod=30 Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.769405 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10d92e7e-6c63-41b9-83d7-663163ee7a00","Type":"ContainerStarted","Data":"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db"} Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.769444 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10d92e7e-6c63-41b9-83d7-663163ee7a00","Type":"ContainerStarted","Data":"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1"} Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.769525 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-metadata" containerID="cri-o://88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db" gracePeriod=30 Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.784890 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.010770428 podStartE2EDuration="6.784857349s" podCreationTimestamp="2025-11-23 07:11:57 +0000 UTC" firstStartedPulling="2025-11-23 07:11:58.859026379 +0000 UTC m=+1334.372417682" lastFinishedPulling="2025-11-23 07:12:02.63311329 +0000 UTC m=+1338.146504603" observedRunningTime="2025-11-23 07:12:03.778105212 +0000 UTC m=+1339.291496535" watchObservedRunningTime="2025-11-23 07:12:03.784857349 +0000 UTC m=+1339.298248662" Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.788399 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad9b24a7-201a-4c2d-8082-8b911f70b76e","Type":"ContainerStarted","Data":"8d072d5b2c6286a0209aa7c28fa33f1ae5c8e9dd5a8c9ebc8d92c4e7a9606d24"} Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.788559 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ad9b24a7-201a-4c2d-8082-8b911f70b76e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8d072d5b2c6286a0209aa7c28fa33f1ae5c8e9dd5a8c9ebc8d92c4e7a9606d24" gracePeriod=30 Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.877527 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.061892026 podStartE2EDuration="6.877499546s" podCreationTimestamp="2025-11-23 07:11:57 +0000 UTC" firstStartedPulling="2025-11-23 07:11:58.823523798 +0000 UTC m=+1334.336915101" lastFinishedPulling="2025-11-23 07:12:02.639131298 +0000 UTC m=+1338.152522621" observedRunningTime="2025-11-23 07:12:03.860266245 +0000 UTC m=+1339.373657548" watchObservedRunningTime="2025-11-23 07:12:03.877499546 +0000 UTC m=+1339.390890839" Nov 23 07:12:03 crc kubenswrapper[4906]: I1123 07:12:03.879975 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.943156716 podStartE2EDuration="6.879966291s" podCreationTimestamp="2025-11-23 07:11:57 +0000 UTC" firstStartedPulling="2025-11-23 07:11:58.694404045 +0000 UTC m=+1334.207795348" lastFinishedPulling="2025-11-23 07:12:02.63121363 +0000 UTC m=+1338.144604923" observedRunningTime="2025-11-23 07:12:03.809521326 +0000 UTC m=+1339.322912649" watchObservedRunningTime="2025-11-23 07:12:03.879966291 +0000 UTC m=+1339.393357594" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.475203 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.517851 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.881024581 podStartE2EDuration="7.517831955s" podCreationTimestamp="2025-11-23 07:11:57 +0000 UTC" firstStartedPulling="2025-11-23 07:11:58.994646313 +0000 UTC m=+1334.508037616" lastFinishedPulling="2025-11-23 07:12:02.631453687 +0000 UTC m=+1338.144844990" observedRunningTime="2025-11-23 07:12:03.889532712 +0000 UTC m=+1339.402924035" watchObservedRunningTime="2025-11-23 07:12:04.517831955 +0000 UTC m=+1340.031223258" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.616730 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-combined-ca-bundle\") pod \"10d92e7e-6c63-41b9-83d7-663163ee7a00\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.616864 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-config-data\") pod \"10d92e7e-6c63-41b9-83d7-663163ee7a00\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.617553 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10d92e7e-6c63-41b9-83d7-663163ee7a00-logs\") pod \"10d92e7e-6c63-41b9-83d7-663163ee7a00\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.618009 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q692w\" (UniqueName: \"kubernetes.io/projected/10d92e7e-6c63-41b9-83d7-663163ee7a00-kube-api-access-q692w\") pod \"10d92e7e-6c63-41b9-83d7-663163ee7a00\" (UID: \"10d92e7e-6c63-41b9-83d7-663163ee7a00\") " Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.618141 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10d92e7e-6c63-41b9-83d7-663163ee7a00-logs" (OuterVolumeSpecName: "logs") pod "10d92e7e-6c63-41b9-83d7-663163ee7a00" (UID: "10d92e7e-6c63-41b9-83d7-663163ee7a00"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.619471 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10d92e7e-6c63-41b9-83d7-663163ee7a00-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.624582 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d92e7e-6c63-41b9-83d7-663163ee7a00-kube-api-access-q692w" (OuterVolumeSpecName: "kube-api-access-q692w") pod "10d92e7e-6c63-41b9-83d7-663163ee7a00" (UID: "10d92e7e-6c63-41b9-83d7-663163ee7a00"). InnerVolumeSpecName "kube-api-access-q692w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.657981 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-config-data" (OuterVolumeSpecName: "config-data") pod "10d92e7e-6c63-41b9-83d7-663163ee7a00" (UID: "10d92e7e-6c63-41b9-83d7-663163ee7a00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.660100 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10d92e7e-6c63-41b9-83d7-663163ee7a00" (UID: "10d92e7e-6c63-41b9-83d7-663163ee7a00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.720990 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q692w\" (UniqueName: \"kubernetes.io/projected/10d92e7e-6c63-41b9-83d7-663163ee7a00-kube-api-access-q692w\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.721035 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.721045 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d92e7e-6c63-41b9-83d7-663163ee7a00-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.804194 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.804207 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10d92e7e-6c63-41b9-83d7-663163ee7a00","Type":"ContainerDied","Data":"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db"} Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.804281 4906 scope.go:117] "RemoveContainer" containerID="88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.804145 4906 generic.go:334] "Generic (PLEG): container finished" podID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerID="88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db" exitCode=0 Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.804500 4906 generic.go:334] "Generic (PLEG): container finished" podID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerID="af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1" exitCode=143 Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.804554 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10d92e7e-6c63-41b9-83d7-663163ee7a00","Type":"ContainerDied","Data":"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1"} Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.804571 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10d92e7e-6c63-41b9-83d7-663163ee7a00","Type":"ContainerDied","Data":"53df4b146e1e01f787ac1466a20848ad875bdf3599de03714bf4557db8547212"} Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.806671 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerStarted","Data":"0ca08c7fe05e0ca1b2007e2e44cbc1096ed71c15967daa351517e9bb8750d5d9"} Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.839840 4906 scope.go:117] "RemoveContainer" containerID="af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.867671 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.881253 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.891637 4906 scope.go:117] "RemoveContainer" containerID="88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db" Nov 23 07:12:04 crc kubenswrapper[4906]: E1123 07:12:04.892240 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db\": container with ID starting with 88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db not found: ID does not exist" containerID="88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.892293 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db"} err="failed to get container status \"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db\": rpc error: code = NotFound desc = could not find container \"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db\": container with ID starting with 88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db not found: ID does not exist" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.892326 4906 scope.go:117] "RemoveContainer" containerID="af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1" Nov 23 07:12:04 crc kubenswrapper[4906]: E1123 07:12:04.892541 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1\": container with ID starting with af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1 not found: ID does not exist" containerID="af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.892577 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1"} err="failed to get container status \"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1\": rpc error: code = NotFound desc = could not find container \"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1\": container with ID starting with af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1 not found: ID does not exist" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.892597 4906 scope.go:117] "RemoveContainer" containerID="88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.893200 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db"} err="failed to get container status \"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db\": rpc error: code = NotFound desc = could not find container \"88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db\": container with ID starting with 88f60c7043dcd13b66efe4d85496c53206586e96b48ad2de8a8a3f323985c0db not found: ID does not exist" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.893225 4906 scope.go:117] "RemoveContainer" containerID="af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.893413 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1"} err="failed to get container status \"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1\": rpc error: code = NotFound desc = could not find container \"af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1\": container with ID starting with af02b6bff56090ab06e93c340c0a244bc49970ba4b3c5729f1cd61c618df8ef1 not found: ID does not exist" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.898931 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:04 crc kubenswrapper[4906]: E1123 07:12:04.899507 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-log" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.899532 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-log" Nov 23 07:12:04 crc kubenswrapper[4906]: E1123 07:12:04.899557 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-metadata" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.899565 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-metadata" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.900056 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-log" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.900097 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" containerName="nova-metadata-metadata" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.901265 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.906327 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.919563 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:12:04 crc kubenswrapper[4906]: I1123 07:12:04.926265 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.041205 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.041306 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2815f02f-7094-4bbd-8699-c1fda30e54eb-logs\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.041348 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.041391 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmkg2\" (UniqueName: \"kubernetes.io/projected/2815f02f-7094-4bbd-8699-c1fda30e54eb-kube-api-access-fmkg2\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.041427 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-config-data\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.143137 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmkg2\" (UniqueName: \"kubernetes.io/projected/2815f02f-7094-4bbd-8699-c1fda30e54eb-kube-api-access-fmkg2\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.143236 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-config-data\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.143335 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.143413 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2815f02f-7094-4bbd-8699-c1fda30e54eb-logs\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.143455 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.144131 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2815f02f-7094-4bbd-8699-c1fda30e54eb-logs\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.147510 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.147648 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.151734 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-config-data\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.164824 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmkg2\" (UniqueName: \"kubernetes.io/projected/2815f02f-7094-4bbd-8699-c1fda30e54eb-kube-api-access-fmkg2\") pod \"nova-metadata-0\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.239461 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.369798 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10d92e7e-6c63-41b9-83d7-663163ee7a00" path="/var/lib/kubelet/pods/10d92e7e-6c63-41b9-83d7-663163ee7a00/volumes" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.808973 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:05 crc kubenswrapper[4906]: W1123 07:12:05.815414 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2815f02f_7094_4bbd_8699_c1fda30e54eb.slice/crio-91527be9609aa30664520e239ad836fa56f7ebae01783de3d3d25969710945cd WatchSource:0}: Error finding container 91527be9609aa30664520e239ad836fa56f7ebae01783de3d3d25969710945cd: Status 404 returned error can't find the container with id 91527be9609aa30664520e239ad836fa56f7ebae01783de3d3d25969710945cd Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.834164 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerStarted","Data":"cada0c8a6e2ce0090307d157d9d2423e1cf703ee5d422589c9268d4381e760b5"} Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.835556 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:12:05 crc kubenswrapper[4906]: I1123 07:12:05.864965 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.442148591 podStartE2EDuration="7.864944133s" podCreationTimestamp="2025-11-23 07:11:58 +0000 UTC" firstStartedPulling="2025-11-23 07:11:59.636353147 +0000 UTC m=+1335.149744440" lastFinishedPulling="2025-11-23 07:12:05.059148679 +0000 UTC m=+1340.572539982" observedRunningTime="2025-11-23 07:12:05.857868477 +0000 UTC m=+1341.371259810" watchObservedRunningTime="2025-11-23 07:12:05.864944133 +0000 UTC m=+1341.378335436" Nov 23 07:12:06 crc kubenswrapper[4906]: I1123 07:12:06.845764 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2815f02f-7094-4bbd-8699-c1fda30e54eb","Type":"ContainerStarted","Data":"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e"} Nov 23 07:12:06 crc kubenswrapper[4906]: I1123 07:12:06.846200 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2815f02f-7094-4bbd-8699-c1fda30e54eb","Type":"ContainerStarted","Data":"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5"} Nov 23 07:12:06 crc kubenswrapper[4906]: I1123 07:12:06.846217 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2815f02f-7094-4bbd-8699-c1fda30e54eb","Type":"ContainerStarted","Data":"91527be9609aa30664520e239ad836fa56f7ebae01783de3d3d25969710945cd"} Nov 23 07:12:06 crc kubenswrapper[4906]: I1123 07:12:06.870164 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.870135791 podStartE2EDuration="2.870135791s" podCreationTimestamp="2025-11-23 07:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:06.86550225 +0000 UTC m=+1342.378893563" watchObservedRunningTime="2025-11-23 07:12:06.870135791 +0000 UTC m=+1342.383527094" Nov 23 07:12:07 crc kubenswrapper[4906]: I1123 07:12:07.862850 4906 generic.go:334] "Generic (PLEG): container finished" podID="35e15932-ac43-4d8e-ba00-481ad71b5816" containerID="6f1bb39f61ceebf4bbb9cb405575f82b1dff5097833cdf28d70e297480072af0" exitCode=0 Nov 23 07:12:07 crc kubenswrapper[4906]: I1123 07:12:07.862914 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h4lvc" event={"ID":"35e15932-ac43-4d8e-ba00-481ad71b5816","Type":"ContainerDied","Data":"6f1bb39f61ceebf4bbb9cb405575f82b1dff5097833cdf28d70e297480072af0"} Nov 23 07:12:07 crc kubenswrapper[4906]: I1123 07:12:07.933085 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 07:12:07 crc kubenswrapper[4906]: I1123 07:12:07.933370 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 07:12:07 crc kubenswrapper[4906]: I1123 07:12:07.971244 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.245854 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.245913 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.357772 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.362844 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.451310 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-d45mm"] Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.451578 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-797bbc649-d45mm" podUID="fee2c65e-0c82-4258-8263-3de044da4402" containerName="dnsmasq-dns" containerID="cri-o://a66239351e06dfa141b579065f17e9ef73f39c529b916a1ac8499620e7915c6d" gracePeriod=10 Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.885072 4906 generic.go:334] "Generic (PLEG): container finished" podID="df30375c-b83c-490c-be3d-2453779a3c06" containerID="96435a9a6698e54650b138b7b8daa6f2fcc2fc3938e89653a935fea62eeb5cf6" exitCode=0 Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.885508 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-54gvc" event={"ID":"df30375c-b83c-490c-be3d-2453779a3c06","Type":"ContainerDied","Data":"96435a9a6698e54650b138b7b8daa6f2fcc2fc3938e89653a935fea62eeb5cf6"} Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.888877 4906 generic.go:334] "Generic (PLEG): container finished" podID="fee2c65e-0c82-4258-8263-3de044da4402" containerID="a66239351e06dfa141b579065f17e9ef73f39c529b916a1ac8499620e7915c6d" exitCode=0 Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.890335 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-d45mm" event={"ID":"fee2c65e-0c82-4258-8263-3de044da4402","Type":"ContainerDied","Data":"a66239351e06dfa141b579065f17e9ef73f39c529b916a1ac8499620e7915c6d"} Nov 23 07:12:08 crc kubenswrapper[4906]: I1123 07:12:08.929526 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.049964 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.135624 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-657rw\" (UniqueName: \"kubernetes.io/projected/fee2c65e-0c82-4258-8263-3de044da4402-kube-api-access-657rw\") pod \"fee2c65e-0c82-4258-8263-3de044da4402\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.135823 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-swift-storage-0\") pod \"fee2c65e-0c82-4258-8263-3de044da4402\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.135914 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-sb\") pod \"fee2c65e-0c82-4258-8263-3de044da4402\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.135989 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-nb\") pod \"fee2c65e-0c82-4258-8263-3de044da4402\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.136046 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-config\") pod \"fee2c65e-0c82-4258-8263-3de044da4402\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.136091 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-svc\") pod \"fee2c65e-0c82-4258-8263-3de044da4402\" (UID: \"fee2c65e-0c82-4258-8263-3de044da4402\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.164949 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee2c65e-0c82-4258-8263-3de044da4402-kube-api-access-657rw" (OuterVolumeSpecName: "kube-api-access-657rw") pod "fee2c65e-0c82-4258-8263-3de044da4402" (UID: "fee2c65e-0c82-4258-8263-3de044da4402"). InnerVolumeSpecName "kube-api-access-657rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.194167 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-config" (OuterVolumeSpecName: "config") pod "fee2c65e-0c82-4258-8263-3de044da4402" (UID: "fee2c65e-0c82-4258-8263-3de044da4402"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.198331 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fee2c65e-0c82-4258-8263-3de044da4402" (UID: "fee2c65e-0c82-4258-8263-3de044da4402"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.199078 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fee2c65e-0c82-4258-8263-3de044da4402" (UID: "fee2c65e-0c82-4258-8263-3de044da4402"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.211143 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fee2c65e-0c82-4258-8263-3de044da4402" (UID: "fee2c65e-0c82-4258-8263-3de044da4402"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.238893 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.238918 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.238930 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-657rw\" (UniqueName: \"kubernetes.io/projected/fee2c65e-0c82-4258-8263-3de044da4402-kube-api-access-657rw\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.238942 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.238951 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.300927 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fee2c65e-0c82-4258-8263-3de044da4402" (UID: "fee2c65e-0c82-4258-8263-3de044da4402"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.335867 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.182:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.335870 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.182:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.341096 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee2c65e-0c82-4258-8263-3de044da4402-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.368646 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.443528 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-combined-ca-bundle\") pod \"35e15932-ac43-4d8e-ba00-481ad71b5816\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.443592 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-config-data\") pod \"35e15932-ac43-4d8e-ba00-481ad71b5816\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.443640 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-scripts\") pod \"35e15932-ac43-4d8e-ba00-481ad71b5816\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.443897 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kf44\" (UniqueName: \"kubernetes.io/projected/35e15932-ac43-4d8e-ba00-481ad71b5816-kube-api-access-6kf44\") pod \"35e15932-ac43-4d8e-ba00-481ad71b5816\" (UID: \"35e15932-ac43-4d8e-ba00-481ad71b5816\") " Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.453574 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-scripts" (OuterVolumeSpecName: "scripts") pod "35e15932-ac43-4d8e-ba00-481ad71b5816" (UID: "35e15932-ac43-4d8e-ba00-481ad71b5816"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.453663 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e15932-ac43-4d8e-ba00-481ad71b5816-kube-api-access-6kf44" (OuterVolumeSpecName: "kube-api-access-6kf44") pod "35e15932-ac43-4d8e-ba00-481ad71b5816" (UID: "35e15932-ac43-4d8e-ba00-481ad71b5816"). InnerVolumeSpecName "kube-api-access-6kf44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.471586 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35e15932-ac43-4d8e-ba00-481ad71b5816" (UID: "35e15932-ac43-4d8e-ba00-481ad71b5816"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.499849 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-config-data" (OuterVolumeSpecName: "config-data") pod "35e15932-ac43-4d8e-ba00-481ad71b5816" (UID: "35e15932-ac43-4d8e-ba00-481ad71b5816"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.546831 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.546869 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.546882 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kf44\" (UniqueName: \"kubernetes.io/projected/35e15932-ac43-4d8e-ba00-481ad71b5816-kube-api-access-6kf44\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.546897 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e15932-ac43-4d8e-ba00-481ad71b5816-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.902141 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-d45mm" event={"ID":"fee2c65e-0c82-4258-8263-3de044da4402","Type":"ContainerDied","Data":"a6074f7996d62eca3feb85117ab5e990011b33df5b80fc761b517f99e148b7e7"} Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.902195 4906 scope.go:117] "RemoveContainer" containerID="a66239351e06dfa141b579065f17e9ef73f39c529b916a1ac8499620e7915c6d" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.902334 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-d45mm" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.907355 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h4lvc" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.910396 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h4lvc" event={"ID":"35e15932-ac43-4d8e-ba00-481ad71b5816","Type":"ContainerDied","Data":"881106b2832b86c7da98f8a1a7b4b1b987139699039512e4a1540be32cb511e1"} Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.910456 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="881106b2832b86c7da98f8a1a7b4b1b987139699039512e4a1540be32cb511e1" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.948928 4906 scope.go:117] "RemoveContainer" containerID="a6009fd7d930fa522902f4cc415e5ac1d70f0ced8343de6d0c4d564d6ae712c0" Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.949748 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-d45mm"] Nov 23 07:12:09 crc kubenswrapper[4906]: I1123 07:12:09.956247 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-d45mm"] Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.102795 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.121224 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.121499 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-log" containerID="cri-o://2846c1bbb7a7f281e8fd181fee816335cda2a4869820e6a6fccd491b87baf0e7" gracePeriod=30 Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.122097 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-api" containerID="cri-o://eea407ed926731b5de4cfa6b5d5d8ba32c637b612478efe45f499b87b1b68fc9" gracePeriod=30 Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.195580 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.196262 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-log" containerID="cri-o://50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5" gracePeriod=30 Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.196821 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-metadata" containerID="cri-o://554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e" gracePeriod=30 Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.248879 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.249188 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.459053 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.594773 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gclt9\" (UniqueName: \"kubernetes.io/projected/df30375c-b83c-490c-be3d-2453779a3c06-kube-api-access-gclt9\") pod \"df30375c-b83c-490c-be3d-2453779a3c06\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.594825 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-scripts\") pod \"df30375c-b83c-490c-be3d-2453779a3c06\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.594914 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-combined-ca-bundle\") pod \"df30375c-b83c-490c-be3d-2453779a3c06\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.595133 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-config-data\") pod \"df30375c-b83c-490c-be3d-2453779a3c06\" (UID: \"df30375c-b83c-490c-be3d-2453779a3c06\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.605539 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df30375c-b83c-490c-be3d-2453779a3c06-kube-api-access-gclt9" (OuterVolumeSpecName: "kube-api-access-gclt9") pod "df30375c-b83c-490c-be3d-2453779a3c06" (UID: "df30375c-b83c-490c-be3d-2453779a3c06"). InnerVolumeSpecName "kube-api-access-gclt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.609807 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-scripts" (OuterVolumeSpecName: "scripts") pod "df30375c-b83c-490c-be3d-2453779a3c06" (UID: "df30375c-b83c-490c-be3d-2453779a3c06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.644884 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-config-data" (OuterVolumeSpecName: "config-data") pod "df30375c-b83c-490c-be3d-2453779a3c06" (UID: "df30375c-b83c-490c-be3d-2453779a3c06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.669769 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df30375c-b83c-490c-be3d-2453779a3c06" (UID: "df30375c-b83c-490c-be3d-2453779a3c06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.697830 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.697868 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.697880 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gclt9\" (UniqueName: \"kubernetes.io/projected/df30375c-b83c-490c-be3d-2453779a3c06-kube-api-access-gclt9\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.697894 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df30375c-b83c-490c-be3d-2453779a3c06-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.762520 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.901698 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-combined-ca-bundle\") pod \"2815f02f-7094-4bbd-8699-c1fda30e54eb\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.901739 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-nova-metadata-tls-certs\") pod \"2815f02f-7094-4bbd-8699-c1fda30e54eb\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.901811 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-config-data\") pod \"2815f02f-7094-4bbd-8699-c1fda30e54eb\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.901946 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmkg2\" (UniqueName: \"kubernetes.io/projected/2815f02f-7094-4bbd-8699-c1fda30e54eb-kube-api-access-fmkg2\") pod \"2815f02f-7094-4bbd-8699-c1fda30e54eb\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.902027 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2815f02f-7094-4bbd-8699-c1fda30e54eb-logs\") pod \"2815f02f-7094-4bbd-8699-c1fda30e54eb\" (UID: \"2815f02f-7094-4bbd-8699-c1fda30e54eb\") " Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.902638 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2815f02f-7094-4bbd-8699-c1fda30e54eb-logs" (OuterVolumeSpecName: "logs") pod "2815f02f-7094-4bbd-8699-c1fda30e54eb" (UID: "2815f02f-7094-4bbd-8699-c1fda30e54eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.920812 4906 generic.go:334] "Generic (PLEG): container finished" podID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerID="2846c1bbb7a7f281e8fd181fee816335cda2a4869820e6a6fccd491b87baf0e7" exitCode=143 Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.920882 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"075c468b-da1c-4e30-b4cf-fe52cb94a76a","Type":"ContainerDied","Data":"2846c1bbb7a7f281e8fd181fee816335cda2a4869820e6a6fccd491b87baf0e7"} Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.924791 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-54gvc" event={"ID":"df30375c-b83c-490c-be3d-2453779a3c06","Type":"ContainerDied","Data":"8fa3b71554771f95e8b74960c97318ac51bb50297d161c7478cddb60493f4998"} Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.924875 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fa3b71554771f95e8b74960c97318ac51bb50297d161c7478cddb60493f4998" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.925011 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-54gvc" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.927064 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2815f02f-7094-4bbd-8699-c1fda30e54eb-kube-api-access-fmkg2" (OuterVolumeSpecName: "kube-api-access-fmkg2") pod "2815f02f-7094-4bbd-8699-c1fda30e54eb" (UID: "2815f02f-7094-4bbd-8699-c1fda30e54eb"). InnerVolumeSpecName "kube-api-access-fmkg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.931221 4906 generic.go:334] "Generic (PLEG): container finished" podID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerID="554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e" exitCode=0 Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.931274 4906 generic.go:334] "Generic (PLEG): container finished" podID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerID="50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5" exitCode=143 Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.931297 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2815f02f-7094-4bbd-8699-c1fda30e54eb","Type":"ContainerDied","Data":"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e"} Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.931356 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2815f02f-7094-4bbd-8699-c1fda30e54eb","Type":"ContainerDied","Data":"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5"} Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.931372 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2815f02f-7094-4bbd-8699-c1fda30e54eb","Type":"ContainerDied","Data":"91527be9609aa30664520e239ad836fa56f7ebae01783de3d3d25969710945cd"} Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.931394 4906 scope.go:117] "RemoveContainer" containerID="554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.931612 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.932010 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-config-data" (OuterVolumeSpecName: "config-data") pod "2815f02f-7094-4bbd-8699-c1fda30e54eb" (UID: "2815f02f-7094-4bbd-8699-c1fda30e54eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.959908 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2815f02f-7094-4bbd-8699-c1fda30e54eb" (UID: "2815f02f-7094-4bbd-8699-c1fda30e54eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.982421 4906 scope.go:117] "RemoveContainer" containerID="50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5" Nov 23 07:12:10 crc kubenswrapper[4906]: I1123 07:12:10.984804 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2815f02f-7094-4bbd-8699-c1fda30e54eb" (UID: "2815f02f-7094-4bbd-8699-c1fda30e54eb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.005073 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmkg2\" (UniqueName: \"kubernetes.io/projected/2815f02f-7094-4bbd-8699-c1fda30e54eb-kube-api-access-fmkg2\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.005116 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2815f02f-7094-4bbd-8699-c1fda30e54eb-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.005127 4906 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.005137 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.005148 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2815f02f-7094-4bbd-8699-c1fda30e54eb-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.021155 4906 scope.go:117] "RemoveContainer" containerID="554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e" Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.025964 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e\": container with ID starting with 554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e not found: ID does not exist" containerID="554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.026024 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e"} err="failed to get container status \"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e\": rpc error: code = NotFound desc = could not find container \"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e\": container with ID starting with 554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e not found: ID does not exist" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.026061 4906 scope.go:117] "RemoveContainer" containerID="50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5" Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.026462 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5\": container with ID starting with 50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5 not found: ID does not exist" containerID="50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.026487 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5"} err="failed to get container status \"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5\": rpc error: code = NotFound desc = could not find container \"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5\": container with ID starting with 50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5 not found: ID does not exist" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.026499 4906 scope.go:117] "RemoveContainer" containerID="554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.026707 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e"} err="failed to get container status \"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e\": rpc error: code = NotFound desc = could not find container \"554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e\": container with ID starting with 554aeef34e8e14a9f3b5e8ae07a66e87c472655bef919ab637337844984c7a4e not found: ID does not exist" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.026726 4906 scope.go:117] "RemoveContainer" containerID="50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.026926 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5"} err="failed to get container status \"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5\": rpc error: code = NotFound desc = could not find container \"50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5\": container with ID starting with 50d7f7d459844816b7611d48197dc21075ae02fb0f3ae7a1e82e42e3e197b7e5 not found: ID does not exist" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.036969 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.037459 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee2c65e-0c82-4258-8263-3de044da4402" containerName="init" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037479 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee2c65e-0c82-4258-8263-3de044da4402" containerName="init" Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.037505 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-log" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037515 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-log" Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.037545 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-metadata" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037553 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-metadata" Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.037566 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee2c65e-0c82-4258-8263-3de044da4402" containerName="dnsmasq-dns" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037573 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee2c65e-0c82-4258-8263-3de044da4402" containerName="dnsmasq-dns" Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.037593 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e15932-ac43-4d8e-ba00-481ad71b5816" containerName="nova-manage" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037601 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e15932-ac43-4d8e-ba00-481ad71b5816" containerName="nova-manage" Nov 23 07:12:11 crc kubenswrapper[4906]: E1123 07:12:11.037630 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df30375c-b83c-490c-be3d-2453779a3c06" containerName="nova-cell1-conductor-db-sync" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037640 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="df30375c-b83c-490c-be3d-2453779a3c06" containerName="nova-cell1-conductor-db-sync" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037849 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee2c65e-0c82-4258-8263-3de044da4402" containerName="dnsmasq-dns" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037864 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-metadata" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037878 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" containerName="nova-metadata-log" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037894 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="df30375c-b83c-490c-be3d-2453779a3c06" containerName="nova-cell1-conductor-db-sync" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.037907 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e15932-ac43-4d8e-ba00-481ad71b5816" containerName="nova-manage" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.038672 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.057004 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.082899 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.110195 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.110291 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h89fv\" (UniqueName: \"kubernetes.io/projected/212ee22b-c3e7-4d9c-b952-5837b5982c63-kube-api-access-h89fv\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.110329 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.212866 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.213131 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h89fv\" (UniqueName: \"kubernetes.io/projected/212ee22b-c3e7-4d9c-b952-5837b5982c63-kube-api-access-h89fv\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.213208 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.217281 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.218174 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.231251 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h89fv\" (UniqueName: \"kubernetes.io/projected/212ee22b-c3e7-4d9c-b952-5837b5982c63-kube-api-access-h89fv\") pod \"nova-cell1-conductor-0\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.269508 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.281242 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.303817 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.305768 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.308066 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.308088 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.316773 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.372081 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2815f02f-7094-4bbd-8699-c1fda30e54eb" path="/var/lib/kubelet/pods/2815f02f-7094-4bbd-8699-c1fda30e54eb/volumes" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.373239 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fee2c65e-0c82-4258-8263-3de044da4402" path="/var/lib/kubelet/pods/fee2c65e-0c82-4258-8263-3de044da4402/volumes" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.382733 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.421872 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.423491 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-config-data\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.423743 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.423804 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-logs\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.423860 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqj9n\" (UniqueName: \"kubernetes.io/projected/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-kube-api-access-cqj9n\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.528146 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.528238 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-config-data\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.528401 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.528429 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-logs\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.528464 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqj9n\" (UniqueName: \"kubernetes.io/projected/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-kube-api-access-cqj9n\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.531455 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-logs\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.539654 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-config-data\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.550268 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.550538 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqj9n\" (UniqueName: \"kubernetes.io/projected/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-kube-api-access-cqj9n\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.551441 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.639611 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.850865 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.953822 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1de8803b-196b-4eae-8b9e-e9dd02e12a7d" containerName="nova-scheduler-scheduler" containerID="cri-o://7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c" gracePeriod=30 Nov 23 07:12:11 crc kubenswrapper[4906]: I1123 07:12:11.954035 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"212ee22b-c3e7-4d9c-b952-5837b5982c63","Type":"ContainerStarted","Data":"7f37ec9bea7956b87da8f1438c4bc7e51543df6e6a3ff0f6230dcb171cf72e7f"} Nov 23 07:12:12 crc kubenswrapper[4906]: I1123 07:12:12.149345 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:12 crc kubenswrapper[4906]: W1123 07:12:12.152626 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5fb5f94_7022_4e27_93a8_1cf08d5ec2fb.slice/crio-82c8a370a4b5fc1d86aad6c3bead2def6fcff91a9426e3a96685873ab31471be WatchSource:0}: Error finding container 82c8a370a4b5fc1d86aad6c3bead2def6fcff91a9426e3a96685873ab31471be: Status 404 returned error can't find the container with id 82c8a370a4b5fc1d86aad6c3bead2def6fcff91a9426e3a96685873ab31471be Nov 23 07:12:12 crc kubenswrapper[4906]: E1123 07:12:12.935467 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:12:12 crc kubenswrapper[4906]: E1123 07:12:12.937114 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:12:12 crc kubenswrapper[4906]: E1123 07:12:12.938668 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:12:12 crc kubenswrapper[4906]: E1123 07:12:12.938759 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1de8803b-196b-4eae-8b9e-e9dd02e12a7d" containerName="nova-scheduler-scheduler" Nov 23 07:12:12 crc kubenswrapper[4906]: I1123 07:12:12.963445 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb","Type":"ContainerStarted","Data":"4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab"} Nov 23 07:12:12 crc kubenswrapper[4906]: I1123 07:12:12.963495 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb","Type":"ContainerStarted","Data":"059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3"} Nov 23 07:12:12 crc kubenswrapper[4906]: I1123 07:12:12.963504 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb","Type":"ContainerStarted","Data":"82c8a370a4b5fc1d86aad6c3bead2def6fcff91a9426e3a96685873ab31471be"} Nov 23 07:12:12 crc kubenswrapper[4906]: I1123 07:12:12.965987 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"212ee22b-c3e7-4d9c-b952-5837b5982c63","Type":"ContainerStarted","Data":"4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef"} Nov 23 07:12:12 crc kubenswrapper[4906]: I1123 07:12:12.966149 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:13 crc kubenswrapper[4906]: I1123 07:12:13.015598 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.015556619 podStartE2EDuration="2.015556619s" podCreationTimestamp="2025-11-23 07:12:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:12.982987415 +0000 UTC m=+1348.496378718" watchObservedRunningTime="2025-11-23 07:12:13.015556619 +0000 UTC m=+1348.528947922" Nov 23 07:12:13 crc kubenswrapper[4906]: I1123 07:12:13.031939 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.031911877 podStartE2EDuration="3.031911877s" podCreationTimestamp="2025-11-23 07:12:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:13.006125901 +0000 UTC m=+1348.519517214" watchObservedRunningTime="2025-11-23 07:12:13.031911877 +0000 UTC m=+1348.545303180" Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.001534 4906 generic.go:334] "Generic (PLEG): container finished" podID="1de8803b-196b-4eae-8b9e-e9dd02e12a7d" containerID="7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c" exitCode=0 Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.001698 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1de8803b-196b-4eae-8b9e-e9dd02e12a7d","Type":"ContainerDied","Data":"7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c"} Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.414523 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.518361 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnr8n\" (UniqueName: \"kubernetes.io/projected/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-kube-api-access-lnr8n\") pod \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.518464 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-combined-ca-bundle\") pod \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.518623 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-config-data\") pod \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\" (UID: \"1de8803b-196b-4eae-8b9e-e9dd02e12a7d\") " Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.526951 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-kube-api-access-lnr8n" (OuterVolumeSpecName: "kube-api-access-lnr8n") pod "1de8803b-196b-4eae-8b9e-e9dd02e12a7d" (UID: "1de8803b-196b-4eae-8b9e-e9dd02e12a7d"). InnerVolumeSpecName "kube-api-access-lnr8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.560293 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1de8803b-196b-4eae-8b9e-e9dd02e12a7d" (UID: "1de8803b-196b-4eae-8b9e-e9dd02e12a7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.568017 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-config-data" (OuterVolumeSpecName: "config-data") pod "1de8803b-196b-4eae-8b9e-e9dd02e12a7d" (UID: "1de8803b-196b-4eae-8b9e-e9dd02e12a7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.621247 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.621293 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnr8n\" (UniqueName: \"kubernetes.io/projected/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-kube-api-access-lnr8n\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:15 crc kubenswrapper[4906]: I1123 07:12:15.621304 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de8803b-196b-4eae-8b9e-e9dd02e12a7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.020199 4906 generic.go:334] "Generic (PLEG): container finished" podID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerID="eea407ed926731b5de4cfa6b5d5d8ba32c637b612478efe45f499b87b1b68fc9" exitCode=0 Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.021204 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"075c468b-da1c-4e30-b4cf-fe52cb94a76a","Type":"ContainerDied","Data":"eea407ed926731b5de4cfa6b5d5d8ba32c637b612478efe45f499b87b1b68fc9"} Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.024106 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1de8803b-196b-4eae-8b9e-e9dd02e12a7d","Type":"ContainerDied","Data":"dacfc102737d204884a1a66c6a0bb44015d08fcb5123629fbfde147482ae46c5"} Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.024222 4906 scope.go:117] "RemoveContainer" containerID="7dda3e79ddfbb826c827c3f0083095450fd9634a370fcc86879b2d69d5fb9b5c" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.024439 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.064182 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.074060 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.096443 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:12:16 crc kubenswrapper[4906]: E1123 07:12:16.098307 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de8803b-196b-4eae-8b9e-e9dd02e12a7d" containerName="nova-scheduler-scheduler" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.098338 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de8803b-196b-4eae-8b9e-e9dd02e12a7d" containerName="nova-scheduler-scheduler" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.098545 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de8803b-196b-4eae-8b9e-e9dd02e12a7d" containerName="nova-scheduler-scheduler" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.099445 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.103205 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.111324 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.232652 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.237585 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc2qn\" (UniqueName: \"kubernetes.io/projected/6d9e0356-8e1b-474e-96af-bb5d2a15954c-kube-api-access-xc2qn\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.237659 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-config-data\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.237805 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.339323 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-combined-ca-bundle\") pod \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.339655 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fmpk\" (UniqueName: \"kubernetes.io/projected/075c468b-da1c-4e30-b4cf-fe52cb94a76a-kube-api-access-5fmpk\") pod \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.340364 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075c468b-da1c-4e30-b4cf-fe52cb94a76a-logs\") pod \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.340426 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-config-data\") pod \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\" (UID: \"075c468b-da1c-4e30-b4cf-fe52cb94a76a\") " Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.340822 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc2qn\" (UniqueName: \"kubernetes.io/projected/6d9e0356-8e1b-474e-96af-bb5d2a15954c-kube-api-access-xc2qn\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.340854 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-config-data\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.340922 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.341275 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/075c468b-da1c-4e30-b4cf-fe52cb94a76a-logs" (OuterVolumeSpecName: "logs") pod "075c468b-da1c-4e30-b4cf-fe52cb94a76a" (UID: "075c468b-da1c-4e30-b4cf-fe52cb94a76a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.346780 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.348197 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/075c468b-da1c-4e30-b4cf-fe52cb94a76a-kube-api-access-5fmpk" (OuterVolumeSpecName: "kube-api-access-5fmpk") pod "075c468b-da1c-4e30-b4cf-fe52cb94a76a" (UID: "075c468b-da1c-4e30-b4cf-fe52cb94a76a"). InnerVolumeSpecName "kube-api-access-5fmpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.348487 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-config-data\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.358970 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc2qn\" (UniqueName: \"kubernetes.io/projected/6d9e0356-8e1b-474e-96af-bb5d2a15954c-kube-api-access-xc2qn\") pod \"nova-scheduler-0\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.376068 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-config-data" (OuterVolumeSpecName: "config-data") pod "075c468b-da1c-4e30-b4cf-fe52cb94a76a" (UID: "075c468b-da1c-4e30-b4cf-fe52cb94a76a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.384007 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "075c468b-da1c-4e30-b4cf-fe52cb94a76a" (UID: "075c468b-da1c-4e30-b4cf-fe52cb94a76a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.449303 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.449424 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fmpk\" (UniqueName: \"kubernetes.io/projected/075c468b-da1c-4e30-b4cf-fe52cb94a76a-kube-api-access-5fmpk\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.449462 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/075c468b-da1c-4e30-b4cf-fe52cb94a76a-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.449495 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075c468b-da1c-4e30-b4cf-fe52cb94a76a-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.580115 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.640732 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:12:16 crc kubenswrapper[4906]: I1123 07:12:16.641850 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.038819 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"075c468b-da1c-4e30-b4cf-fe52cb94a76a","Type":"ContainerDied","Data":"62ed8be502476a60d6da1cd8eb02485b292dbcbe09ad30956472bde7e69c2894"} Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.038916 4906 scope.go:117] "RemoveContainer" containerID="eea407ed926731b5de4cfa6b5d5d8ba32c637b612478efe45f499b87b1b68fc9" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.038858 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.094431 4906 scope.go:117] "RemoveContainer" containerID="2846c1bbb7a7f281e8fd181fee816335cda2a4869820e6a6fccd491b87baf0e7" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.103547 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.138531 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.166725 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.178310 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:17 crc kubenswrapper[4906]: E1123 07:12:17.179148 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-api" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.179187 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-api" Nov 23 07:12:17 crc kubenswrapper[4906]: E1123 07:12:17.179219 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-log" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.179258 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-log" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.179852 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-log" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.179890 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" containerName="nova-api-api" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.181810 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.188491 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.198420 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.272085 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ldxl\" (UniqueName: \"kubernetes.io/projected/41907c13-3210-4498-af54-89fc144beb12-kube-api-access-9ldxl\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.272153 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.272182 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41907c13-3210-4498-af54-89fc144beb12-logs\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.272553 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.375205 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41907c13-3210-4498-af54-89fc144beb12-logs\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.375554 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.375819 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ldxl\" (UniqueName: \"kubernetes.io/projected/41907c13-3210-4498-af54-89fc144beb12-kube-api-access-9ldxl\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.375928 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.379174 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41907c13-3210-4498-af54-89fc144beb12-logs\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.381414 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="075c468b-da1c-4e30-b4cf-fe52cb94a76a" path="/var/lib/kubelet/pods/075c468b-da1c-4e30-b4cf-fe52cb94a76a/volumes" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.383826 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de8803b-196b-4eae-8b9e-e9dd02e12a7d" path="/var/lib/kubelet/pods/1de8803b-196b-4eae-8b9e-e9dd02e12a7d/volumes" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.384419 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.386984 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.399458 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ldxl\" (UniqueName: \"kubernetes.io/projected/41907c13-3210-4498-af54-89fc144beb12-kube-api-access-9ldxl\") pod \"nova-api-0\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " pod="openstack/nova-api-0" Nov 23 07:12:17 crc kubenswrapper[4906]: I1123 07:12:17.646768 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:18 crc kubenswrapper[4906]: I1123 07:12:18.056907 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d9e0356-8e1b-474e-96af-bb5d2a15954c","Type":"ContainerStarted","Data":"4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af"} Nov 23 07:12:18 crc kubenswrapper[4906]: I1123 07:12:18.057567 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d9e0356-8e1b-474e-96af-bb5d2a15954c","Type":"ContainerStarted","Data":"cc79759fb4ba71ccda1de2526dfb443445643f202458c594fa322c366271a158"} Nov 23 07:12:18 crc kubenswrapper[4906]: I1123 07:12:18.090120 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.090085295 podStartE2EDuration="2.090085295s" podCreationTimestamp="2025-11-23 07:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:18.077474775 +0000 UTC m=+1353.590866088" watchObservedRunningTime="2025-11-23 07:12:18.090085295 +0000 UTC m=+1353.603476608" Nov 23 07:12:18 crc kubenswrapper[4906]: I1123 07:12:18.225442 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:19 crc kubenswrapper[4906]: I1123 07:12:19.069829 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41907c13-3210-4498-af54-89fc144beb12","Type":"ContainerStarted","Data":"35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8"} Nov 23 07:12:19 crc kubenswrapper[4906]: I1123 07:12:19.070865 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41907c13-3210-4498-af54-89fc144beb12","Type":"ContainerStarted","Data":"957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5"} Nov 23 07:12:19 crc kubenswrapper[4906]: I1123 07:12:19.070902 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41907c13-3210-4498-af54-89fc144beb12","Type":"ContainerStarted","Data":"5623817a7d77cee4a416f95570f6e2e9922c91433c9c101e4ff2b6f114b480d2"} Nov 23 07:12:19 crc kubenswrapper[4906]: I1123 07:12:19.096909 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.096887196 podStartE2EDuration="2.096887196s" podCreationTimestamp="2025-11-23 07:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:19.094149604 +0000 UTC m=+1354.607540917" watchObservedRunningTime="2025-11-23 07:12:19.096887196 +0000 UTC m=+1354.610278499" Nov 23 07:12:21 crc kubenswrapper[4906]: I1123 07:12:21.427740 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 23 07:12:21 crc kubenswrapper[4906]: I1123 07:12:21.581053 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 07:12:21 crc kubenswrapper[4906]: I1123 07:12:21.640278 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:12:21 crc kubenswrapper[4906]: I1123 07:12:21.640360 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:12:22 crc kubenswrapper[4906]: I1123 07:12:22.655038 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:12:22 crc kubenswrapper[4906]: I1123 07:12:22.655078 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:12:26 crc kubenswrapper[4906]: I1123 07:12:26.580915 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 07:12:26 crc kubenswrapper[4906]: I1123 07:12:26.628364 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 07:12:27 crc kubenswrapper[4906]: I1123 07:12:27.226747 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 07:12:27 crc kubenswrapper[4906]: I1123 07:12:27.647653 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:12:27 crc kubenswrapper[4906]: I1123 07:12:27.647732 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:12:28 crc kubenswrapper[4906]: I1123 07:12:28.730083 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:12:28 crc kubenswrapper[4906]: I1123 07:12:28.730654 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:12:29 crc kubenswrapper[4906]: I1123 07:12:29.045848 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 07:12:31 crc kubenswrapper[4906]: I1123 07:12:31.649044 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:12:31 crc kubenswrapper[4906]: I1123 07:12:31.651935 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:12:31 crc kubenswrapper[4906]: I1123 07:12:31.655962 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:12:32 crc kubenswrapper[4906]: I1123 07:12:32.245032 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.258592 4906 generic.go:334] "Generic (PLEG): container finished" podID="ad9b24a7-201a-4c2d-8082-8b911f70b76e" containerID="8d072d5b2c6286a0209aa7c28fa33f1ae5c8e9dd5a8c9ebc8d92c4e7a9606d24" exitCode=137 Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.258815 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad9b24a7-201a-4c2d-8082-8b911f70b76e","Type":"ContainerDied","Data":"8d072d5b2c6286a0209aa7c28fa33f1ae5c8e9dd5a8c9ebc8d92c4e7a9606d24"} Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.367957 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.533606 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-combined-ca-bundle\") pod \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.534002 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-config-data\") pod \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.534098 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g5hx\" (UniqueName: \"kubernetes.io/projected/ad9b24a7-201a-4c2d-8082-8b911f70b76e-kube-api-access-2g5hx\") pod \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\" (UID: \"ad9b24a7-201a-4c2d-8082-8b911f70b76e\") " Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.543784 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad9b24a7-201a-4c2d-8082-8b911f70b76e-kube-api-access-2g5hx" (OuterVolumeSpecName: "kube-api-access-2g5hx") pod "ad9b24a7-201a-4c2d-8082-8b911f70b76e" (UID: "ad9b24a7-201a-4c2d-8082-8b911f70b76e"). InnerVolumeSpecName "kube-api-access-2g5hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.573502 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-config-data" (OuterVolumeSpecName: "config-data") pod "ad9b24a7-201a-4c2d-8082-8b911f70b76e" (UID: "ad9b24a7-201a-4c2d-8082-8b911f70b76e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.596774 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad9b24a7-201a-4c2d-8082-8b911f70b76e" (UID: "ad9b24a7-201a-4c2d-8082-8b911f70b76e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.636561 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.636592 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g5hx\" (UniqueName: \"kubernetes.io/projected/ad9b24a7-201a-4c2d-8082-8b911f70b76e-kube-api-access-2g5hx\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:34 crc kubenswrapper[4906]: I1123 07:12:34.636607 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b24a7-201a-4c2d-8082-8b911f70b76e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.275914 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ad9b24a7-201a-4c2d-8082-8b911f70b76e","Type":"ContainerDied","Data":"e1ebce5afce804edbb82b4b6a67f7646dfd885bbdf590363528b40fae34ff0fe"} Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.276039 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.276151 4906 scope.go:117] "RemoveContainer" containerID="8d072d5b2c6286a0209aa7c28fa33f1ae5c8e9dd5a8c9ebc8d92c4e7a9606d24" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.346236 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.398152 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.398250 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:12:35 crc kubenswrapper[4906]: E1123 07:12:35.401092 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9b24a7-201a-4c2d-8082-8b911f70b76e" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.401157 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9b24a7-201a-4c2d-8082-8b911f70b76e" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.401797 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9b24a7-201a-4c2d-8082-8b911f70b76e" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.403363 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.413052 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.413487 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.413773 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.414176 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.559965 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.560483 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.560525 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.560841 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mp7r\" (UniqueName: \"kubernetes.io/projected/6cee328a-b746-486d-850b-148ca1e66eea-kube-api-access-4mp7r\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.560901 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.663049 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.663200 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mp7r\" (UniqueName: \"kubernetes.io/projected/6cee328a-b746-486d-850b-148ca1e66eea-kube-api-access-4mp7r\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.663225 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.663268 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.663636 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.667595 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.668094 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.670023 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.673180 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.685794 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mp7r\" (UniqueName: \"kubernetes.io/projected/6cee328a-b746-486d-850b-148ca1e66eea-kube-api-access-4mp7r\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:35 crc kubenswrapper[4906]: I1123 07:12:35.730075 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:36 crc kubenswrapper[4906]: I1123 07:12:36.184983 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:12:36 crc kubenswrapper[4906]: I1123 07:12:36.288563 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cee328a-b746-486d-850b-148ca1e66eea","Type":"ContainerStarted","Data":"0defa47ee7111204fd4f55204c7ca6efa62d6a16a0d4e3e0329a19e00784a029"} Nov 23 07:12:37 crc kubenswrapper[4906]: I1123 07:12:37.300045 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cee328a-b746-486d-850b-148ca1e66eea","Type":"ContainerStarted","Data":"18d1eb931158c01bd8771394295d1f098c58d8673d4374a70905fb56a217df1f"} Nov 23 07:12:37 crc kubenswrapper[4906]: I1123 07:12:37.326151 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.326121139 podStartE2EDuration="2.326121139s" podCreationTimestamp="2025-11-23 07:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:37.319109465 +0000 UTC m=+1372.832500808" watchObservedRunningTime="2025-11-23 07:12:37.326121139 +0000 UTC m=+1372.839512442" Nov 23 07:12:37 crc kubenswrapper[4906]: I1123 07:12:37.370823 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad9b24a7-201a-4c2d-8082-8b911f70b76e" path="/var/lib/kubelet/pods/ad9b24a7-201a-4c2d-8082-8b911f70b76e/volumes" Nov 23 07:12:37 crc kubenswrapper[4906]: I1123 07:12:37.651961 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:12:37 crc kubenswrapper[4906]: I1123 07:12:37.652498 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:12:37 crc kubenswrapper[4906]: I1123 07:12:37.653835 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:12:37 crc kubenswrapper[4906]: I1123 07:12:37.657022 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.311247 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.314898 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.513955 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-944h6"] Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.515880 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.544727 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-944h6"] Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.633046 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxxmj\" (UniqueName: \"kubernetes.io/projected/d40394f2-6a98-49df-813a-17db4ced9cd2-kube-api-access-wxxmj\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.633127 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.633172 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.633233 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.633260 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.633295 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-config\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.734467 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.734773 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.734801 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-config\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.734893 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxxmj\" (UniqueName: \"kubernetes.io/projected/d40394f2-6a98-49df-813a-17db4ced9cd2-kube-api-access-wxxmj\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.734928 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.734964 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.735956 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.735993 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.736005 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.736345 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-config\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.737143 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.770914 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxxmj\" (UniqueName: \"kubernetes.io/projected/d40394f2-6a98-49df-813a-17db4ced9cd2-kube-api-access-wxxmj\") pod \"dnsmasq-dns-5d7f54fb65-944h6\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:38 crc kubenswrapper[4906]: I1123 07:12:38.846059 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:39 crc kubenswrapper[4906]: I1123 07:12:39.385018 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-944h6"] Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.337104 4906 generic.go:334] "Generic (PLEG): container finished" podID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerID="a22234cdf3c895d8bd75acf98c4f78605b788dfa3db4b1b07efc828a4980c561" exitCode=0 Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.338662 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" event={"ID":"d40394f2-6a98-49df-813a-17db4ced9cd2","Type":"ContainerDied","Data":"a22234cdf3c895d8bd75acf98c4f78605b788dfa3db4b1b07efc828a4980c561"} Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.338714 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" event={"ID":"d40394f2-6a98-49df-813a-17db4ced9cd2","Type":"ContainerStarted","Data":"efd52e7666be2f2f4da0cc1d0490f9a3f79d80bbbd90cea212bdde53a11e6688"} Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.731039 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.957953 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.958316 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-central-agent" containerID="cri-o://ce47dd2e8dc264da48a333d312b4f60fdd8a13b1b6573e5fc8dd87fa49c1e4a7" gracePeriod=30 Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.958416 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="proxy-httpd" containerID="cri-o://cada0c8a6e2ce0090307d157d9d2423e1cf703ee5d422589c9268d4381e760b5" gracePeriod=30 Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.958481 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="sg-core" containerID="cri-o://0ca08c7fe05e0ca1b2007e2e44cbc1096ed71c15967daa351517e9bb8750d5d9" gracePeriod=30 Nov 23 07:12:40 crc kubenswrapper[4906]: I1123 07:12:40.958522 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-notification-agent" containerID="cri-o://1f4258c9e792706d58c0124470407407625159d5bd411557fb8aa27bf93252e5" gracePeriod=30 Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.353281 4906 generic.go:334] "Generic (PLEG): container finished" podID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerID="cada0c8a6e2ce0090307d157d9d2423e1cf703ee5d422589c9268d4381e760b5" exitCode=0 Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.353859 4906 generic.go:334] "Generic (PLEG): container finished" podID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerID="0ca08c7fe05e0ca1b2007e2e44cbc1096ed71c15967daa351517e9bb8750d5d9" exitCode=2 Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.353354 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerDied","Data":"cada0c8a6e2ce0090307d157d9d2423e1cf703ee5d422589c9268d4381e760b5"} Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.353924 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerDied","Data":"0ca08c7fe05e0ca1b2007e2e44cbc1096ed71c15967daa351517e9bb8750d5d9"} Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.376020 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.376080 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" event={"ID":"d40394f2-6a98-49df-813a-17db4ced9cd2","Type":"ContainerStarted","Data":"9310cf8e52799cd9ffd94d046c01ec36de9b4fb21b64eff04510a5c86bb2a2cd"} Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.404439 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" podStartSLOduration=3.404404931 podStartE2EDuration="3.404404931s" podCreationTimestamp="2025-11-23 07:12:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:41.385397714 +0000 UTC m=+1376.898789027" watchObservedRunningTime="2025-11-23 07:12:41.404404931 +0000 UTC m=+1376.917796244" Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.724757 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.725074 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-log" containerID="cri-o://957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5" gracePeriod=30 Nov 23 07:12:41 crc kubenswrapper[4906]: I1123 07:12:41.725206 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-api" containerID="cri-o://35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8" gracePeriod=30 Nov 23 07:12:42 crc kubenswrapper[4906]: I1123 07:12:42.371693 4906 generic.go:334] "Generic (PLEG): container finished" podID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerID="ce47dd2e8dc264da48a333d312b4f60fdd8a13b1b6573e5fc8dd87fa49c1e4a7" exitCode=0 Nov 23 07:12:42 crc kubenswrapper[4906]: I1123 07:12:42.371960 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerDied","Data":"ce47dd2e8dc264da48a333d312b4f60fdd8a13b1b6573e5fc8dd87fa49c1e4a7"} Nov 23 07:12:42 crc kubenswrapper[4906]: I1123 07:12:42.374843 4906 generic.go:334] "Generic (PLEG): container finished" podID="41907c13-3210-4498-af54-89fc144beb12" containerID="957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5" exitCode=143 Nov 23 07:12:42 crc kubenswrapper[4906]: I1123 07:12:42.375002 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41907c13-3210-4498-af54-89fc144beb12","Type":"ContainerDied","Data":"957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5"} Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.425222 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.429586 4906 generic.go:334] "Generic (PLEG): container finished" podID="41907c13-3210-4498-af54-89fc144beb12" containerID="35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8" exitCode=0 Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.429957 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41907c13-3210-4498-af54-89fc144beb12","Type":"ContainerDied","Data":"35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8"} Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.430027 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41907c13-3210-4498-af54-89fc144beb12","Type":"ContainerDied","Data":"5623817a7d77cee4a416f95570f6e2e9922c91433c9c101e4ff2b6f114b480d2"} Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.430053 4906 scope.go:117] "RemoveContainer" containerID="35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.531350 4906 scope.go:117] "RemoveContainer" containerID="957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.536230 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-combined-ca-bundle\") pod \"41907c13-3210-4498-af54-89fc144beb12\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.536321 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41907c13-3210-4498-af54-89fc144beb12-logs\") pod \"41907c13-3210-4498-af54-89fc144beb12\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.536551 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ldxl\" (UniqueName: \"kubernetes.io/projected/41907c13-3210-4498-af54-89fc144beb12-kube-api-access-9ldxl\") pod \"41907c13-3210-4498-af54-89fc144beb12\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.536586 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data\") pod \"41907c13-3210-4498-af54-89fc144beb12\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.537158 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41907c13-3210-4498-af54-89fc144beb12-logs" (OuterVolumeSpecName: "logs") pod "41907c13-3210-4498-af54-89fc144beb12" (UID: "41907c13-3210-4498-af54-89fc144beb12"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.568747 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41907c13-3210-4498-af54-89fc144beb12-kube-api-access-9ldxl" (OuterVolumeSpecName: "kube-api-access-9ldxl") pod "41907c13-3210-4498-af54-89fc144beb12" (UID: "41907c13-3210-4498-af54-89fc144beb12"). InnerVolumeSpecName "kube-api-access-9ldxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:45 crc kubenswrapper[4906]: E1123 07:12:45.583114 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data podName:41907c13-3210-4498-af54-89fc144beb12 nodeName:}" failed. No retries permitted until 2025-11-23 07:12:46.083066354 +0000 UTC m=+1381.596457657 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data") pod "41907c13-3210-4498-af54-89fc144beb12" (UID: "41907c13-3210-4498-af54-89fc144beb12") : error deleting /var/lib/kubelet/pods/41907c13-3210-4498-af54-89fc144beb12/volume-subpaths: remove /var/lib/kubelet/pods/41907c13-3210-4498-af54-89fc144beb12/volume-subpaths: no such file or directory Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.587896 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41907c13-3210-4498-af54-89fc144beb12" (UID: "41907c13-3210-4498-af54-89fc144beb12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.639564 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ldxl\" (UniqueName: \"kubernetes.io/projected/41907c13-3210-4498-af54-89fc144beb12-kube-api-access-9ldxl\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.639961 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.639971 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41907c13-3210-4498-af54-89fc144beb12-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.650999 4906 scope.go:117] "RemoveContainer" containerID="35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8" Nov 23 07:12:45 crc kubenswrapper[4906]: E1123 07:12:45.651635 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8\": container with ID starting with 35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8 not found: ID does not exist" containerID="35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.651712 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8"} err="failed to get container status \"35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8\": rpc error: code = NotFound desc = could not find container \"35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8\": container with ID starting with 35c88fd58d8bd74da313a0774b43a34d569ed01e0ef827b0ac794859b0eb85a8 not found: ID does not exist" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.651746 4906 scope.go:117] "RemoveContainer" containerID="957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5" Nov 23 07:12:45 crc kubenswrapper[4906]: E1123 07:12:45.652705 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5\": container with ID starting with 957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5 not found: ID does not exist" containerID="957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.652744 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5"} err="failed to get container status \"957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5\": rpc error: code = NotFound desc = could not find container \"957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5\": container with ID starting with 957fdc7561260e0140b0be1c96f6deca47524d51c8c332d057b507e2b8f10fd5 not found: ID does not exist" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.730531 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:45 crc kubenswrapper[4906]: I1123 07:12:45.755850 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.150028 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data\") pod \"41907c13-3210-4498-af54-89fc144beb12\" (UID: \"41907c13-3210-4498-af54-89fc144beb12\") " Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.159907 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data" (OuterVolumeSpecName: "config-data") pod "41907c13-3210-4498-af54-89fc144beb12" (UID: "41907c13-3210-4498-af54-89fc144beb12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.255293 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41907c13-3210-4498-af54-89fc144beb12-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.453991 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.472800 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.495307 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.512181 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.531187 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:46 crc kubenswrapper[4906]: E1123 07:12:46.532046 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-api" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.532067 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-api" Nov 23 07:12:46 crc kubenswrapper[4906]: E1123 07:12:46.532096 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-log" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.532104 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-log" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.532291 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-log" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.532319 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="41907c13-3210-4498-af54-89fc144beb12" containerName="nova-api-api" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.533465 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.536756 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.536938 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.543089 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.543825 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.664109 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-public-tls-certs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.664167 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q52n6\" (UniqueName: \"kubernetes.io/projected/49b156ab-cd4c-4787-b489-f834c93c4bdc-kube-api-access-q52n6\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.664235 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-config-data\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.664437 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.664833 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49b156ab-cd4c-4787-b489-f834c93c4bdc-logs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.664993 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.733773 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-flw9z"] Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.735109 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.737382 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.742533 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.758057 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-flw9z"] Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.767188 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49b156ab-cd4c-4787-b489-f834c93c4bdc-logs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.767275 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.767319 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-public-tls-certs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.767343 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q52n6\" (UniqueName: \"kubernetes.io/projected/49b156ab-cd4c-4787-b489-f834c93c4bdc-kube-api-access-q52n6\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.767375 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-config-data\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.767410 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.768276 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49b156ab-cd4c-4787-b489-f834c93c4bdc-logs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.773364 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.774503 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-public-tls-certs\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.793416 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.793822 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-config-data\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.797207 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q52n6\" (UniqueName: \"kubernetes.io/projected/49b156ab-cd4c-4787-b489-f834c93c4bdc-kube-api-access-q52n6\") pod \"nova-api-0\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.855086 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.869069 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-scripts\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.869131 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rmj\" (UniqueName: \"kubernetes.io/projected/7a2c54f8-cc51-4e33-9b70-62107c26f267-kube-api-access-88rmj\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.869163 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-config-data\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.869187 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.972649 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-scripts\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.973240 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rmj\" (UniqueName: \"kubernetes.io/projected/7a2c54f8-cc51-4e33-9b70-62107c26f267-kube-api-access-88rmj\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.973280 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-config-data\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.973309 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.979385 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-scripts\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.980747 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-config-data\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.984980 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:46 crc kubenswrapper[4906]: I1123 07:12:46.990455 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rmj\" (UniqueName: \"kubernetes.io/projected/7a2c54f8-cc51-4e33-9b70-62107c26f267-kube-api-access-88rmj\") pod \"nova-cell1-cell-mapping-flw9z\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.056156 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.346554 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:47 crc kubenswrapper[4906]: W1123 07:12:47.365655 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49b156ab_cd4c_4787_b489_f834c93c4bdc.slice/crio-2ffbbbf8f74866c61d2cacd7f694f8eb572c72badce43b1f52728db705e50663 WatchSource:0}: Error finding container 2ffbbbf8f74866c61d2cacd7f694f8eb572c72badce43b1f52728db705e50663: Status 404 returned error can't find the container with id 2ffbbbf8f74866c61d2cacd7f694f8eb572c72badce43b1f52728db705e50663 Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.373541 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41907c13-3210-4498-af54-89fc144beb12" path="/var/lib/kubelet/pods/41907c13-3210-4498-af54-89fc144beb12/volumes" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.471017 4906 generic.go:334] "Generic (PLEG): container finished" podID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerID="1f4258c9e792706d58c0124470407407625159d5bd411557fb8aa27bf93252e5" exitCode=0 Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.471105 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerDied","Data":"1f4258c9e792706d58c0124470407407625159d5bd411557fb8aa27bf93252e5"} Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.473440 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49b156ab-cd4c-4787-b489-f834c93c4bdc","Type":"ContainerStarted","Data":"2ffbbbf8f74866c61d2cacd7f694f8eb572c72badce43b1f52728db705e50663"} Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.682656 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.682766 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-flw9z"] Nov 23 07:12:47 crc kubenswrapper[4906]: W1123 07:12:47.688332 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a2c54f8_cc51_4e33_9b70_62107c26f267.slice/crio-a8689dad3dfe0dee7db99e03adc77188230fb43f132814c743e9b43acf4d1ff5 WatchSource:0}: Error finding container a8689dad3dfe0dee7db99e03adc77188230fb43f132814c743e9b43acf4d1ff5: Status 404 returned error can't find the container with id a8689dad3dfe0dee7db99e03adc77188230fb43f132814c743e9b43acf4d1ff5 Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.789661 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-scripts\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.789768 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-ceilometer-tls-certs\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.789825 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-config-data\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.789854 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-run-httpd\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.790014 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-sg-core-conf-yaml\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.790043 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-combined-ca-bundle\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.790070 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-log-httpd\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.790118 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4sw6\" (UniqueName: \"kubernetes.io/projected/a91aaae7-e5f2-4f11-b3e4-a15179305476-kube-api-access-j4sw6\") pod \"a91aaae7-e5f2-4f11-b3e4-a15179305476\" (UID: \"a91aaae7-e5f2-4f11-b3e4-a15179305476\") " Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.793130 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.798307 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.801031 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a91aaae7-e5f2-4f11-b3e4-a15179305476-kube-api-access-j4sw6" (OuterVolumeSpecName: "kube-api-access-j4sw6") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "kube-api-access-j4sw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.808747 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-scripts" (OuterVolumeSpecName: "scripts") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.849661 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.864061 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.892141 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.892166 4906 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.892175 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a91aaae7-e5f2-4f11-b3e4-a15179305476-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.892184 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4sw6\" (UniqueName: \"kubernetes.io/projected/a91aaae7-e5f2-4f11-b3e4-a15179305476-kube-api-access-j4sw6\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.892194 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.892203 4906 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.929083 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.975238 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-config-data" (OuterVolumeSpecName: "config-data") pod "a91aaae7-e5f2-4f11-b3e4-a15179305476" (UID: "a91aaae7-e5f2-4f11-b3e4-a15179305476"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.994564 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:47 crc kubenswrapper[4906]: I1123 07:12:47.994612 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a91aaae7-e5f2-4f11-b3e4-a15179305476-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.512662 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49b156ab-cd4c-4787-b489-f834c93c4bdc","Type":"ContainerStarted","Data":"2f7ed5f961b2bd0f59e6b02d62ce9827f0984c4d42db858f675dc1f9f6502ef0"} Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.513045 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49b156ab-cd4c-4787-b489-f834c93c4bdc","Type":"ContainerStarted","Data":"bdc434bd0f33ceed31b2e69f77c9f48db3077f2f569b169bd1b3ef43e453a62b"} Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.515065 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flw9z" event={"ID":"7a2c54f8-cc51-4e33-9b70-62107c26f267","Type":"ContainerStarted","Data":"f1bcaee5f685e06a5e0f095d05465381185a3d0e78e940373d9a346353454dc6"} Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.515097 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flw9z" event={"ID":"7a2c54f8-cc51-4e33-9b70-62107c26f267","Type":"ContainerStarted","Data":"a8689dad3dfe0dee7db99e03adc77188230fb43f132814c743e9b43acf4d1ff5"} Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.519779 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a91aaae7-e5f2-4f11-b3e4-a15179305476","Type":"ContainerDied","Data":"1b35615d4f7e05c5cfb605d405a85cc02c37f8ad0235c0aab47be954cbf06368"} Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.519849 4906 scope.go:117] "RemoveContainer" containerID="cada0c8a6e2ce0090307d157d9d2423e1cf703ee5d422589c9268d4381e760b5" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.520226 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.543942 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.543913366 podStartE2EDuration="2.543913366s" podCreationTimestamp="2025-11-23 07:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:48.540984799 +0000 UTC m=+1384.054376102" watchObservedRunningTime="2025-11-23 07:12:48.543913366 +0000 UTC m=+1384.057304679" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.571648 4906 scope.go:117] "RemoveContainer" containerID="0ca08c7fe05e0ca1b2007e2e44cbc1096ed71c15967daa351517e9bb8750d5d9" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.572883 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-flw9z" podStartSLOduration=2.572859694 podStartE2EDuration="2.572859694s" podCreationTimestamp="2025-11-23 07:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:48.563409977 +0000 UTC m=+1384.076801280" watchObservedRunningTime="2025-11-23 07:12:48.572859694 +0000 UTC m=+1384.086251007" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.617971 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.618472 4906 scope.go:117] "RemoveContainer" containerID="1f4258c9e792706d58c0124470407407625159d5bd411557fb8aa27bf93252e5" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.634011 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.642556 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:12:48 crc kubenswrapper[4906]: E1123 07:12:48.643169 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-notification-agent" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643191 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-notification-agent" Nov 23 07:12:48 crc kubenswrapper[4906]: E1123 07:12:48.643220 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="sg-core" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643227 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="sg-core" Nov 23 07:12:48 crc kubenswrapper[4906]: E1123 07:12:48.643248 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="proxy-httpd" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643255 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="proxy-httpd" Nov 23 07:12:48 crc kubenswrapper[4906]: E1123 07:12:48.643272 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-central-agent" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643278 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-central-agent" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643479 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-notification-agent" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643503 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="sg-core" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643514 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="proxy-httpd" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.643523 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" containerName="ceilometer-central-agent" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.645615 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.648453 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.649994 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.656972 4906 scope.go:117] "RemoveContainer" containerID="ce47dd2e8dc264da48a333d312b4f60fdd8a13b1b6573e5fc8dd87fa49c1e4a7" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.657473 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.657888 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.717879 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.717940 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.717987 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-scripts\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.718033 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.718090 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-run-httpd\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.718696 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-log-httpd\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.718997 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-config-data\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.719055 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nqq6\" (UniqueName: \"kubernetes.io/projected/79f707e8-3c2b-4541-94f9-b7cc09cdda72-kube-api-access-7nqq6\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.821552 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-log-httpd\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.821731 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-config-data\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.821798 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nqq6\" (UniqueName: \"kubernetes.io/projected/79f707e8-3c2b-4541-94f9-b7cc09cdda72-kube-api-access-7nqq6\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.821870 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.821903 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.821980 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-scripts\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.822045 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.822120 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-run-httpd\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.822125 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-log-httpd\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.822862 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-run-httpd\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.828875 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-config-data\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.830085 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-scripts\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.830390 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.832126 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.841750 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.843343 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nqq6\" (UniqueName: \"kubernetes.io/projected/79f707e8-3c2b-4541-94f9-b7cc09cdda72-kube-api-access-7nqq6\") pod \"ceilometer-0\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " pod="openstack/ceilometer-0" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.849136 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.952207 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-lwpwn"] Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.952528 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" podUID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerName="dnsmasq-dns" containerID="cri-o://9cf15debcaa784ebb7c366d5efc1dfc07119d5f901f8b3e1db3f4dfeef24c333" gracePeriod=10 Nov 23 07:12:48 crc kubenswrapper[4906]: I1123 07:12:48.974399 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.376736 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a91aaae7-e5f2-4f11-b3e4-a15179305476" path="/var/lib/kubelet/pods/a91aaae7-e5f2-4f11-b3e4-a15179305476/volumes" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.508213 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.509516 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.562298 4906 generic.go:334] "Generic (PLEG): container finished" podID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerID="9cf15debcaa784ebb7c366d5efc1dfc07119d5f901f8b3e1db3f4dfeef24c333" exitCode=0 Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.562379 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" event={"ID":"009ce17b-acb9-47b7-b4ed-6ebca0ef635a","Type":"ContainerDied","Data":"9cf15debcaa784ebb7c366d5efc1dfc07119d5f901f8b3e1db3f4dfeef24c333"} Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.562441 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" event={"ID":"009ce17b-acb9-47b7-b4ed-6ebca0ef635a","Type":"ContainerDied","Data":"128a8304c71f69c84f55fbdf15e6046ac6779a39aaf7b2808a4fd17d9d694a33"} Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.562456 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="128a8304c71f69c84f55fbdf15e6046ac6779a39aaf7b2808a4fd17d9d694a33" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.564214 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerStarted","Data":"0dbb56eb073b8b5727506abf95dffd56c860f95e0f51c45ee769559f6796acb3"} Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.592385 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.645880 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-config\") pod \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.645959 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-swift-storage-0\") pod \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.646057 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-nb\") pod \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.646172 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjrs9\" (UniqueName: \"kubernetes.io/projected/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-kube-api-access-mjrs9\") pod \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.646205 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-sb\") pod \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.646293 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-svc\") pod \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\" (UID: \"009ce17b-acb9-47b7-b4ed-6ebca0ef635a\") " Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.666810 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-kube-api-access-mjrs9" (OuterVolumeSpecName: "kube-api-access-mjrs9") pod "009ce17b-acb9-47b7-b4ed-6ebca0ef635a" (UID: "009ce17b-acb9-47b7-b4ed-6ebca0ef635a"). InnerVolumeSpecName "kube-api-access-mjrs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.707894 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "009ce17b-acb9-47b7-b4ed-6ebca0ef635a" (UID: "009ce17b-acb9-47b7-b4ed-6ebca0ef635a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.713453 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-config" (OuterVolumeSpecName: "config") pod "009ce17b-acb9-47b7-b4ed-6ebca0ef635a" (UID: "009ce17b-acb9-47b7-b4ed-6ebca0ef635a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.719077 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "009ce17b-acb9-47b7-b4ed-6ebca0ef635a" (UID: "009ce17b-acb9-47b7-b4ed-6ebca0ef635a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.719106 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "009ce17b-acb9-47b7-b4ed-6ebca0ef635a" (UID: "009ce17b-acb9-47b7-b4ed-6ebca0ef635a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.738430 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "009ce17b-acb9-47b7-b4ed-6ebca0ef635a" (UID: "009ce17b-acb9-47b7-b4ed-6ebca0ef635a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.749407 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjrs9\" (UniqueName: \"kubernetes.io/projected/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-kube-api-access-mjrs9\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.749762 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.749831 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.749887 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.749952 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:49 crc kubenswrapper[4906]: I1123 07:12:49.750012 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/009ce17b-acb9-47b7-b4ed-6ebca0ef635a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:50 crc kubenswrapper[4906]: I1123 07:12:50.575351 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerStarted","Data":"f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df"} Nov 23 07:12:50 crc kubenswrapper[4906]: I1123 07:12:50.575398 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-lwpwn" Nov 23 07:12:50 crc kubenswrapper[4906]: I1123 07:12:50.682723 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-lwpwn"] Nov 23 07:12:50 crc kubenswrapper[4906]: I1123 07:12:50.691756 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-lwpwn"] Nov 23 07:12:51 crc kubenswrapper[4906]: I1123 07:12:51.376051 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" path="/var/lib/kubelet/pods/009ce17b-acb9-47b7-b4ed-6ebca0ef635a/volumes" Nov 23 07:12:51 crc kubenswrapper[4906]: I1123 07:12:51.588641 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerStarted","Data":"009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5"} Nov 23 07:12:51 crc kubenswrapper[4906]: I1123 07:12:51.589344 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerStarted","Data":"f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f"} Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.146709 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8r2d9"] Nov 23 07:12:53 crc kubenswrapper[4906]: E1123 07:12:53.147580 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerName="init" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.147592 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerName="init" Nov 23 07:12:53 crc kubenswrapper[4906]: E1123 07:12:53.147609 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerName="dnsmasq-dns" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.147616 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerName="dnsmasq-dns" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.147863 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="009ce17b-acb9-47b7-b4ed-6ebca0ef635a" containerName="dnsmasq-dns" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.149281 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.174381 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8r2d9"] Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.245256 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-utilities\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.245299 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-catalog-content\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.245571 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skg5c\" (UniqueName: \"kubernetes.io/projected/c2432dda-e7b8-4e36-823c-fa1a3075b57b-kube-api-access-skg5c\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.347592 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-utilities\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.347651 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-catalog-content\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.347755 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skg5c\" (UniqueName: \"kubernetes.io/projected/c2432dda-e7b8-4e36-823c-fa1a3075b57b-kube-api-access-skg5c\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.348259 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-utilities\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.348373 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-catalog-content\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.374086 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skg5c\" (UniqueName: \"kubernetes.io/projected/c2432dda-e7b8-4e36-823c-fa1a3075b57b-kube-api-access-skg5c\") pod \"redhat-operators-8r2d9\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.483200 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.641651 4906 generic.go:334] "Generic (PLEG): container finished" podID="7a2c54f8-cc51-4e33-9b70-62107c26f267" containerID="f1bcaee5f685e06a5e0f095d05465381185a3d0e78e940373d9a346353454dc6" exitCode=0 Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.641725 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flw9z" event={"ID":"7a2c54f8-cc51-4e33-9b70-62107c26f267","Type":"ContainerDied","Data":"f1bcaee5f685e06a5e0f095d05465381185a3d0e78e940373d9a346353454dc6"} Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.657505 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerStarted","Data":"4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0"} Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.658139 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:12:53 crc kubenswrapper[4906]: I1123 07:12:53.697405 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8302791750000003 podStartE2EDuration="5.697374121s" podCreationTimestamp="2025-11-23 07:12:48 +0000 UTC" firstStartedPulling="2025-11-23 07:12:49.509270791 +0000 UTC m=+1385.022662094" lastFinishedPulling="2025-11-23 07:12:52.376365737 +0000 UTC m=+1387.889757040" observedRunningTime="2025-11-23 07:12:53.691587939 +0000 UTC m=+1389.204979242" watchObservedRunningTime="2025-11-23 07:12:53.697374121 +0000 UTC m=+1389.210765414" Nov 23 07:12:54 crc kubenswrapper[4906]: I1123 07:12:54.024576 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8r2d9"] Nov 23 07:12:54 crc kubenswrapper[4906]: W1123 07:12:54.034472 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2432dda_e7b8_4e36_823c_fa1a3075b57b.slice/crio-b85b4dd29efddcd027497a37381c56b53e43347c81ad7291ba1209607e17f70d WatchSource:0}: Error finding container b85b4dd29efddcd027497a37381c56b53e43347c81ad7291ba1209607e17f70d: Status 404 returned error can't find the container with id b85b4dd29efddcd027497a37381c56b53e43347c81ad7291ba1209607e17f70d Nov 23 07:12:54 crc kubenswrapper[4906]: I1123 07:12:54.669307 4906 generic.go:334] "Generic (PLEG): container finished" podID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerID="14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4" exitCode=0 Nov 23 07:12:54 crc kubenswrapper[4906]: I1123 07:12:54.669427 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r2d9" event={"ID":"c2432dda-e7b8-4e36-823c-fa1a3075b57b","Type":"ContainerDied","Data":"14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4"} Nov 23 07:12:54 crc kubenswrapper[4906]: I1123 07:12:54.669730 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r2d9" event={"ID":"c2432dda-e7b8-4e36-823c-fa1a3075b57b","Type":"ContainerStarted","Data":"b85b4dd29efddcd027497a37381c56b53e43347c81ad7291ba1209607e17f70d"} Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.096186 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.195648 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-combined-ca-bundle\") pod \"7a2c54f8-cc51-4e33-9b70-62107c26f267\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.195762 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-config-data\") pod \"7a2c54f8-cc51-4e33-9b70-62107c26f267\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.195825 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88rmj\" (UniqueName: \"kubernetes.io/projected/7a2c54f8-cc51-4e33-9b70-62107c26f267-kube-api-access-88rmj\") pod \"7a2c54f8-cc51-4e33-9b70-62107c26f267\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.195859 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-scripts\") pod \"7a2c54f8-cc51-4e33-9b70-62107c26f267\" (UID: \"7a2c54f8-cc51-4e33-9b70-62107c26f267\") " Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.202282 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-scripts" (OuterVolumeSpecName: "scripts") pod "7a2c54f8-cc51-4e33-9b70-62107c26f267" (UID: "7a2c54f8-cc51-4e33-9b70-62107c26f267"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.219073 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a2c54f8-cc51-4e33-9b70-62107c26f267-kube-api-access-88rmj" (OuterVolumeSpecName: "kube-api-access-88rmj") pod "7a2c54f8-cc51-4e33-9b70-62107c26f267" (UID: "7a2c54f8-cc51-4e33-9b70-62107c26f267"). InnerVolumeSpecName "kube-api-access-88rmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.228872 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-config-data" (OuterVolumeSpecName: "config-data") pod "7a2c54f8-cc51-4e33-9b70-62107c26f267" (UID: "7a2c54f8-cc51-4e33-9b70-62107c26f267"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.230979 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a2c54f8-cc51-4e33-9b70-62107c26f267" (UID: "7a2c54f8-cc51-4e33-9b70-62107c26f267"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.297745 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.297782 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.297791 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88rmj\" (UniqueName: \"kubernetes.io/projected/7a2c54f8-cc51-4e33-9b70-62107c26f267-kube-api-access-88rmj\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.297804 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a2c54f8-cc51-4e33-9b70-62107c26f267-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.682350 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flw9z" event={"ID":"7a2c54f8-cc51-4e33-9b70-62107c26f267","Type":"ContainerDied","Data":"a8689dad3dfe0dee7db99e03adc77188230fb43f132814c743e9b43acf4d1ff5"} Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.682400 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8689dad3dfe0dee7db99e03adc77188230fb43f132814c743e9b43acf4d1ff5" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.682365 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flw9z" Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.684120 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r2d9" event={"ID":"c2432dda-e7b8-4e36-823c-fa1a3075b57b","Type":"ContainerStarted","Data":"f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7"} Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.851138 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.851396 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6d9e0356-8e1b-474e-96af-bb5d2a15954c" containerName="nova-scheduler-scheduler" containerID="cri-o://4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" gracePeriod=30 Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.871918 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.873517 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-api" containerID="cri-o://2f7ed5f961b2bd0f59e6b02d62ce9827f0984c4d42db858f675dc1f9f6502ef0" gracePeriod=30 Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.873589 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-log" containerID="cri-o://bdc434bd0f33ceed31b2e69f77c9f48db3077f2f569b169bd1b3ef43e453a62b" gracePeriod=30 Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.957423 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.957787 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-metadata" containerID="cri-o://4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab" gracePeriod=30 Nov 23 07:12:55 crc kubenswrapper[4906]: I1123 07:12:55.957754 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-log" containerID="cri-o://059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3" gracePeriod=30 Nov 23 07:12:56 crc kubenswrapper[4906]: E1123 07:12:56.583608 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:12:56 crc kubenswrapper[4906]: E1123 07:12:56.585178 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:12:56 crc kubenswrapper[4906]: E1123 07:12:56.587012 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:12:56 crc kubenswrapper[4906]: E1123 07:12:56.587087 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6d9e0356-8e1b-474e-96af-bb5d2a15954c" containerName="nova-scheduler-scheduler" Nov 23 07:12:56 crc kubenswrapper[4906]: I1123 07:12:56.707351 4906 generic.go:334] "Generic (PLEG): container finished" podID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerID="059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3" exitCode=143 Nov 23 07:12:56 crc kubenswrapper[4906]: I1123 07:12:56.707723 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb","Type":"ContainerDied","Data":"059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3"} Nov 23 07:12:56 crc kubenswrapper[4906]: I1123 07:12:56.712328 4906 generic.go:334] "Generic (PLEG): container finished" podID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerID="2f7ed5f961b2bd0f59e6b02d62ce9827f0984c4d42db858f675dc1f9f6502ef0" exitCode=0 Nov 23 07:12:56 crc kubenswrapper[4906]: I1123 07:12:56.712362 4906 generic.go:334] "Generic (PLEG): container finished" podID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerID="bdc434bd0f33ceed31b2e69f77c9f48db3077f2f569b169bd1b3ef43e453a62b" exitCode=143 Nov 23 07:12:56 crc kubenswrapper[4906]: I1123 07:12:56.712382 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49b156ab-cd4c-4787-b489-f834c93c4bdc","Type":"ContainerDied","Data":"2f7ed5f961b2bd0f59e6b02d62ce9827f0984c4d42db858f675dc1f9f6502ef0"} Nov 23 07:12:56 crc kubenswrapper[4906]: I1123 07:12:56.712414 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49b156ab-cd4c-4787-b489-f834c93c4bdc","Type":"ContainerDied","Data":"bdc434bd0f33ceed31b2e69f77c9f48db3077f2f569b169bd1b3ef43e453a62b"} Nov 23 07:12:56 crc kubenswrapper[4906]: I1123 07:12:56.957761 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.042952 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-internal-tls-certs\") pod \"49b156ab-cd4c-4787-b489-f834c93c4bdc\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.043026 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-config-data\") pod \"49b156ab-cd4c-4787-b489-f834c93c4bdc\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.043176 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49b156ab-cd4c-4787-b489-f834c93c4bdc-logs\") pod \"49b156ab-cd4c-4787-b489-f834c93c4bdc\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.043215 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-public-tls-certs\") pod \"49b156ab-cd4c-4787-b489-f834c93c4bdc\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.043292 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-combined-ca-bundle\") pod \"49b156ab-cd4c-4787-b489-f834c93c4bdc\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.043337 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q52n6\" (UniqueName: \"kubernetes.io/projected/49b156ab-cd4c-4787-b489-f834c93c4bdc-kube-api-access-q52n6\") pod \"49b156ab-cd4c-4787-b489-f834c93c4bdc\" (UID: \"49b156ab-cd4c-4787-b489-f834c93c4bdc\") " Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.043652 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49b156ab-cd4c-4787-b489-f834c93c4bdc-logs" (OuterVolumeSpecName: "logs") pod "49b156ab-cd4c-4787-b489-f834c93c4bdc" (UID: "49b156ab-cd4c-4787-b489-f834c93c4bdc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.043775 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49b156ab-cd4c-4787-b489-f834c93c4bdc-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.060336 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b156ab-cd4c-4787-b489-f834c93c4bdc-kube-api-access-q52n6" (OuterVolumeSpecName: "kube-api-access-q52n6") pod "49b156ab-cd4c-4787-b489-f834c93c4bdc" (UID: "49b156ab-cd4c-4787-b489-f834c93c4bdc"). InnerVolumeSpecName "kube-api-access-q52n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.076041 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-config-data" (OuterVolumeSpecName: "config-data") pod "49b156ab-cd4c-4787-b489-f834c93c4bdc" (UID: "49b156ab-cd4c-4787-b489-f834c93c4bdc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.080128 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49b156ab-cd4c-4787-b489-f834c93c4bdc" (UID: "49b156ab-cd4c-4787-b489-f834c93c4bdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.119884 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "49b156ab-cd4c-4787-b489-f834c93c4bdc" (UID: "49b156ab-cd4c-4787-b489-f834c93c4bdc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.120716 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "49b156ab-cd4c-4787-b489-f834c93c4bdc" (UID: "49b156ab-cd4c-4787-b489-f834c93c4bdc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.147251 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.147308 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.147329 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q52n6\" (UniqueName: \"kubernetes.io/projected/49b156ab-cd4c-4787-b489-f834c93c4bdc-kube-api-access-q52n6\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.147350 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.147370 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49b156ab-cd4c-4787-b489-f834c93c4bdc-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.724821 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.724788 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"49b156ab-cd4c-4787-b489-f834c93c4bdc","Type":"ContainerDied","Data":"2ffbbbf8f74866c61d2cacd7f694f8eb572c72badce43b1f52728db705e50663"} Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.724981 4906 scope.go:117] "RemoveContainer" containerID="2f7ed5f961b2bd0f59e6b02d62ce9827f0984c4d42db858f675dc1f9f6502ef0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.727865 4906 generic.go:334] "Generic (PLEG): container finished" podID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerID="f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7" exitCode=0 Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.727903 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r2d9" event={"ID":"c2432dda-e7b8-4e36-823c-fa1a3075b57b","Type":"ContainerDied","Data":"f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7"} Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.754416 4906 scope.go:117] "RemoveContainer" containerID="bdc434bd0f33ceed31b2e69f77c9f48db3077f2f569b169bd1b3ef43e453a62b" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.788987 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.798454 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.844860 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:57 crc kubenswrapper[4906]: E1123 07:12:57.845466 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2c54f8-cc51-4e33-9b70-62107c26f267" containerName="nova-manage" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.845485 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2c54f8-cc51-4e33-9b70-62107c26f267" containerName="nova-manage" Nov 23 07:12:57 crc kubenswrapper[4906]: E1123 07:12:57.845507 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-api" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.845514 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-api" Nov 23 07:12:57 crc kubenswrapper[4906]: E1123 07:12:57.845562 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-log" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.845569 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-log" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.847189 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2c54f8-cc51-4e33-9b70-62107c26f267" containerName="nova-manage" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.847317 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-api" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.847362 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" containerName="nova-api-log" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.852744 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.855448 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.856453 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.856558 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.870386 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.975971 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-config-data\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.976024 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6jg6\" (UniqueName: \"kubernetes.io/projected/1d7b2888-eaed-4879-9f60-e4264aed2287-kube-api-access-v6jg6\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.976044 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7b2888-eaed-4879-9f60-e4264aed2287-logs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.976071 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.976583 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:57 crc kubenswrapper[4906]: I1123 07:12:57.976779 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.079163 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-config-data\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.079216 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6jg6\" (UniqueName: \"kubernetes.io/projected/1d7b2888-eaed-4879-9f60-e4264aed2287-kube-api-access-v6jg6\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.079237 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7b2888-eaed-4879-9f60-e4264aed2287-logs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.079286 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.079347 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.079375 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.079788 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7b2888-eaed-4879-9f60-e4264aed2287-logs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.086760 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-public-tls-certs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.088173 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.101310 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.105316 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-config-data\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.109220 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6jg6\" (UniqueName: \"kubernetes.io/projected/1d7b2888-eaed-4879-9f60-e4264aed2287-kube-api-access-v6jg6\") pod \"nova-api-0\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.179391 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.669478 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:12:58 crc kubenswrapper[4906]: W1123 07:12:58.672434 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d7b2888_eaed_4879_9f60_e4264aed2287.slice/crio-44520f7b636e7cc8da583f7ba7d9f3522a3d30b934dfeafc72f7c6f3464440a6 WatchSource:0}: Error finding container 44520f7b636e7cc8da583f7ba7d9f3522a3d30b934dfeafc72f7c6f3464440a6: Status 404 returned error can't find the container with id 44520f7b636e7cc8da583f7ba7d9f3522a3d30b934dfeafc72f7c6f3464440a6 Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.740929 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r2d9" event={"ID":"c2432dda-e7b8-4e36-823c-fa1a3075b57b","Type":"ContainerStarted","Data":"da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8"} Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.745568 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d7b2888-eaed-4879-9f60-e4264aed2287","Type":"ContainerStarted","Data":"44520f7b636e7cc8da583f7ba7d9f3522a3d30b934dfeafc72f7c6f3464440a6"} Nov 23 07:12:58 crc kubenswrapper[4906]: I1123 07:12:58.766401 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8r2d9" podStartSLOduration=2.240164896 podStartE2EDuration="5.766378372s" podCreationTimestamp="2025-11-23 07:12:53 +0000 UTC" firstStartedPulling="2025-11-23 07:12:54.676975569 +0000 UTC m=+1390.190366862" lastFinishedPulling="2025-11-23 07:12:58.203189045 +0000 UTC m=+1393.716580338" observedRunningTime="2025-11-23 07:12:58.762547282 +0000 UTC m=+1394.275938595" watchObservedRunningTime="2025-11-23 07:12:58.766378372 +0000 UTC m=+1394.279769675" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.111058 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8pttz"] Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.113094 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.127048 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8pttz"] Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.174134 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:44484->10.217.0.189:8775: read: connection reset by peer" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.174134 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:44494->10.217.0.189:8775: read: connection reset by peer" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.204653 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-utilities\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.204738 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-catalog-content\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.204952 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fprc\" (UniqueName: \"kubernetes.io/projected/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-kube-api-access-9fprc\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.306810 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-utilities\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.307236 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-catalog-content\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.307313 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fprc\" (UniqueName: \"kubernetes.io/projected/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-kube-api-access-9fprc\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.307383 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-utilities\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.307813 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-catalog-content\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.326448 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fprc\" (UniqueName: \"kubernetes.io/projected/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-kube-api-access-9fprc\") pod \"community-operators-8pttz\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.369905 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49b156ab-cd4c-4787-b489-f834c93c4bdc" path="/var/lib/kubelet/pods/49b156ab-cd4c-4787-b489-f834c93c4bdc/volumes" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.484532 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.650414 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.716016 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-config-data\") pod \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.716084 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-combined-ca-bundle\") pod \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.716109 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-logs\") pod \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.716200 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqj9n\" (UniqueName: \"kubernetes.io/projected/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-kube-api-access-cqj9n\") pod \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.716243 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-nova-metadata-tls-certs\") pod \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\" (UID: \"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb\") " Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.718189 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-logs" (OuterVolumeSpecName: "logs") pod "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" (UID: "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.731028 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-kube-api-access-cqj9n" (OuterVolumeSpecName: "kube-api-access-cqj9n") pod "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" (UID: "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb"). InnerVolumeSpecName "kube-api-access-cqj9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.772035 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-config-data" (OuterVolumeSpecName: "config-data") pod "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" (UID: "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.778257 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" (UID: "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.783075 4906 generic.go:334] "Generic (PLEG): container finished" podID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerID="4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab" exitCode=0 Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.783144 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb","Type":"ContainerDied","Data":"4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab"} Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.783175 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb","Type":"ContainerDied","Data":"82c8a370a4b5fc1d86aad6c3bead2def6fcff91a9426e3a96685873ab31471be"} Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.783196 4906 scope.go:117] "RemoveContainer" containerID="4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.783345 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.795995 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d7b2888-eaed-4879-9f60-e4264aed2287","Type":"ContainerStarted","Data":"77ce16e41b99ad023af4aa1620f90f3b6cbe24953a476cead47e6c790b48478b"} Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.796034 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d7b2888-eaed-4879-9f60-e4264aed2287","Type":"ContainerStarted","Data":"d7113d47184e50909b0d719000d22c051debd5b0dca75d3f957a7612dc3670a1"} Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.818341 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.818375 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.818385 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqj9n\" (UniqueName: \"kubernetes.io/projected/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-kube-api-access-cqj9n\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.818394 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.831296 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.831275076 podStartE2EDuration="2.831275076s" podCreationTimestamp="2025-11-23 07:12:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:12:59.815194354 +0000 UTC m=+1395.328585657" watchObservedRunningTime="2025-11-23 07:12:59.831275076 +0000 UTC m=+1395.344666379" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.834965 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" (UID: "a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.839359 4906 scope.go:117] "RemoveContainer" containerID="059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.859728 4906 scope.go:117] "RemoveContainer" containerID="4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab" Nov 23 07:12:59 crc kubenswrapper[4906]: E1123 07:12:59.860374 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab\": container with ID starting with 4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab not found: ID does not exist" containerID="4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.860427 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab"} err="failed to get container status \"4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab\": rpc error: code = NotFound desc = could not find container \"4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab\": container with ID starting with 4b5ff2f179849849589309e245cf862075b187685e3ad6fe24a278f21203f7ab not found: ID does not exist" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.860460 4906 scope.go:117] "RemoveContainer" containerID="059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3" Nov 23 07:12:59 crc kubenswrapper[4906]: E1123 07:12:59.860858 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3\": container with ID starting with 059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3 not found: ID does not exist" containerID="059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.860890 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3"} err="failed to get container status \"059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3\": rpc error: code = NotFound desc = could not find container \"059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3\": container with ID starting with 059592571a9de5c925b94072a4f3d485a5147b89c712396a40862a66d9010ba3 not found: ID does not exist" Nov 23 07:12:59 crc kubenswrapper[4906]: I1123 07:12:59.919914 4906 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:00 crc kubenswrapper[4906]: W1123 07:13:00.012915 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadeef82e_07c0_45b4_a3c7_0ea6fcd790dd.slice/crio-a5da67ebcb43f72804afba14957ff2601d59462a0739ad7d044da4f610882f0c WatchSource:0}: Error finding container a5da67ebcb43f72804afba14957ff2601d59462a0739ad7d044da4f610882f0c: Status 404 returned error can't find the container with id a5da67ebcb43f72804afba14957ff2601d59462a0739ad7d044da4f610882f0c Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.017758 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8pttz"] Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.211604 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.222562 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.243148 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:00 crc kubenswrapper[4906]: E1123 07:13:00.243626 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-metadata" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.243645 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-metadata" Nov 23 07:13:00 crc kubenswrapper[4906]: E1123 07:13:00.243688 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-log" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.243697 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-log" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.243875 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-metadata" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.243902 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" containerName="nova-metadata-log" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.244932 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.252654 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.277028 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.277876 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.327422 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.327492 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-config-data\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.327529 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqpwt\" (UniqueName: \"kubernetes.io/projected/00d152f2-1543-4b14-b214-73747a3f9a26-kube-api-access-xqpwt\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.327871 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00d152f2-1543-4b14-b214-73747a3f9a26-logs\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.327998 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.430364 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.430488 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.430524 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-config-data\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.430547 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqpwt\" (UniqueName: \"kubernetes.io/projected/00d152f2-1543-4b14-b214-73747a3f9a26-kube-api-access-xqpwt\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.430702 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00d152f2-1543-4b14-b214-73747a3f9a26-logs\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.431382 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00d152f2-1543-4b14-b214-73747a3f9a26-logs\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.438834 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-config-data\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.439181 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.441592 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.449177 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqpwt\" (UniqueName: \"kubernetes.io/projected/00d152f2-1543-4b14-b214-73747a3f9a26-kube-api-access-xqpwt\") pod \"nova-metadata-0\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.596667 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:13:00 crc kubenswrapper[4906]: E1123 07:13:00.802449 4906 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d9e0356_8e1b_474e_96af_bb5d2a15954c.slice/crio-conmon-4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af.scope\": RecentStats: unable to find data in memory cache]" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.806393 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.814249 4906 generic.go:334] "Generic (PLEG): container finished" podID="6d9e0356-8e1b-474e-96af-bb5d2a15954c" containerID="4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" exitCode=0 Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.814362 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.814573 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d9e0356-8e1b-474e-96af-bb5d2a15954c","Type":"ContainerDied","Data":"4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af"} Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.814604 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d9e0356-8e1b-474e-96af-bb5d2a15954c","Type":"ContainerDied","Data":"cc79759fb4ba71ccda1de2526dfb443445643f202458c594fa322c366271a158"} Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.814624 4906 scope.go:117] "RemoveContainer" containerID="4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.817503 4906 generic.go:334] "Generic (PLEG): container finished" podID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerID="28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc" exitCode=0 Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.817613 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pttz" event={"ID":"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd","Type":"ContainerDied","Data":"28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc"} Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.817638 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pttz" event={"ID":"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd","Type":"ContainerStarted","Data":"a5da67ebcb43f72804afba14957ff2601d59462a0739ad7d044da4f610882f0c"} Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.840544 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-combined-ca-bundle\") pod \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.840770 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc2qn\" (UniqueName: \"kubernetes.io/projected/6d9e0356-8e1b-474e-96af-bb5d2a15954c-kube-api-access-xc2qn\") pod \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.840892 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-config-data\") pod \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\" (UID: \"6d9e0356-8e1b-474e-96af-bb5d2a15954c\") " Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.853549 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9e0356-8e1b-474e-96af-bb5d2a15954c-kube-api-access-xc2qn" (OuterVolumeSpecName: "kube-api-access-xc2qn") pod "6d9e0356-8e1b-474e-96af-bb5d2a15954c" (UID: "6d9e0356-8e1b-474e-96af-bb5d2a15954c"). InnerVolumeSpecName "kube-api-access-xc2qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.869186 4906 scope.go:117] "RemoveContainer" containerID="4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" Nov 23 07:13:00 crc kubenswrapper[4906]: E1123 07:13:00.870195 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af\": container with ID starting with 4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af not found: ID does not exist" containerID="4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.870259 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af"} err="failed to get container status \"4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af\": rpc error: code = NotFound desc = could not find container \"4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af\": container with ID starting with 4f343724a05f153a17e4125dbe58cd606da550526f21e9c5e13302cc949c98af not found: ID does not exist" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.886754 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-config-data" (OuterVolumeSpecName: "config-data") pod "6d9e0356-8e1b-474e-96af-bb5d2a15954c" (UID: "6d9e0356-8e1b-474e-96af-bb5d2a15954c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:00 crc kubenswrapper[4906]: I1123 07:13:00.901291 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d9e0356-8e1b-474e-96af-bb5d2a15954c" (UID: "6d9e0356-8e1b-474e-96af-bb5d2a15954c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:00.944350 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:00.944382 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc2qn\" (UniqueName: \"kubernetes.io/projected/6d9e0356-8e1b-474e-96af-bb5d2a15954c-kube-api-access-xc2qn\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:00.944393 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9e0356-8e1b-474e-96af-bb5d2a15954c-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.157136 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.317930 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.327867 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.337375 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:01 crc kubenswrapper[4906]: E1123 07:13:01.337917 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9e0356-8e1b-474e-96af-bb5d2a15954c" containerName="nova-scheduler-scheduler" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.337933 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9e0356-8e1b-474e-96af-bb5d2a15954c" containerName="nova-scheduler-scheduler" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.338146 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9e0356-8e1b-474e-96af-bb5d2a15954c" containerName="nova-scheduler-scheduler" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.338956 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.344493 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.362610 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.406854 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9e0356-8e1b-474e-96af-bb5d2a15954c" path="/var/lib/kubelet/pods/6d9e0356-8e1b-474e-96af-bb5d2a15954c/volumes" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.407720 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb" path="/var/lib/kubelet/pods/a5fb5f94-7022-4e27-93a8-1cf08d5ec2fb/volumes" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.465632 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-config-data\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.465695 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmwrz\" (UniqueName: \"kubernetes.io/projected/da74c07b-2832-4d8e-9508-415fdc3cac71-kube-api-access-pmwrz\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.465771 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.568001 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-config-data\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.568057 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmwrz\" (UniqueName: \"kubernetes.io/projected/da74c07b-2832-4d8e-9508-415fdc3cac71-kube-api-access-pmwrz\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.568117 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.575717 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.578592 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-config-data\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.585168 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmwrz\" (UniqueName: \"kubernetes.io/projected/da74c07b-2832-4d8e-9508-415fdc3cac71-kube-api-access-pmwrz\") pod \"nova-scheduler-0\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.672431 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.852550 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00d152f2-1543-4b14-b214-73747a3f9a26","Type":"ContainerStarted","Data":"59399848d4fd8c548b0ad9fe61db4277a707a18507241aca88683dc29eebb13a"} Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.852993 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00d152f2-1543-4b14-b214-73747a3f9a26","Type":"ContainerStarted","Data":"cfe0d8e3ca16e113f2016b61cb50f6b5b415f288a07bd395effcd5f6002c7ed6"} Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.853007 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00d152f2-1543-4b14-b214-73747a3f9a26","Type":"ContainerStarted","Data":"c1ba7e6a6a3f0bc8878932b83d10ea4c8acb3eabbbf1c12aed17e0000d862edf"} Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.859093 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pttz" event={"ID":"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd","Type":"ContainerStarted","Data":"0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6"} Nov 23 07:13:01 crc kubenswrapper[4906]: I1123 07:13:01.887207 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.8871798659999999 podStartE2EDuration="1.887179866s" podCreationTimestamp="2025-11-23 07:13:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:13:01.870998662 +0000 UTC m=+1397.384389965" watchObservedRunningTime="2025-11-23 07:13:01.887179866 +0000 UTC m=+1397.400571179" Nov 23 07:13:02 crc kubenswrapper[4906]: I1123 07:13:02.209820 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:02 crc kubenswrapper[4906]: W1123 07:13:02.217826 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda74c07b_2832_4d8e_9508_415fdc3cac71.slice/crio-26285d7556a44fb9a5857cf96431aae6fce54b8ddebef74d066f8afbf62a94b5 WatchSource:0}: Error finding container 26285d7556a44fb9a5857cf96431aae6fce54b8ddebef74d066f8afbf62a94b5: Status 404 returned error can't find the container with id 26285d7556a44fb9a5857cf96431aae6fce54b8ddebef74d066f8afbf62a94b5 Nov 23 07:13:02 crc kubenswrapper[4906]: I1123 07:13:02.872128 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da74c07b-2832-4d8e-9508-415fdc3cac71","Type":"ContainerStarted","Data":"4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f"} Nov 23 07:13:02 crc kubenswrapper[4906]: I1123 07:13:02.872544 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da74c07b-2832-4d8e-9508-415fdc3cac71","Type":"ContainerStarted","Data":"26285d7556a44fb9a5857cf96431aae6fce54b8ddebef74d066f8afbf62a94b5"} Nov 23 07:13:02 crc kubenswrapper[4906]: I1123 07:13:02.874661 4906 generic.go:334] "Generic (PLEG): container finished" podID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerID="0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6" exitCode=0 Nov 23 07:13:02 crc kubenswrapper[4906]: I1123 07:13:02.875006 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pttz" event={"ID":"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd","Type":"ContainerDied","Data":"0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6"} Nov 23 07:13:02 crc kubenswrapper[4906]: I1123 07:13:02.900262 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.90023542 podStartE2EDuration="1.90023542s" podCreationTimestamp="2025-11-23 07:13:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:13:02.889901279 +0000 UTC m=+1398.403292592" watchObservedRunningTime="2025-11-23 07:13:02.90023542 +0000 UTC m=+1398.413626733" Nov 23 07:13:03 crc kubenswrapper[4906]: I1123 07:13:03.483718 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:13:03 crc kubenswrapper[4906]: I1123 07:13:03.483785 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:13:03 crc kubenswrapper[4906]: I1123 07:13:03.894050 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pttz" event={"ID":"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd","Type":"ContainerStarted","Data":"46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321"} Nov 23 07:13:03 crc kubenswrapper[4906]: I1123 07:13:03.916319 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8pttz" podStartSLOduration=2.259158491 podStartE2EDuration="4.916295604s" podCreationTimestamp="2025-11-23 07:12:59 +0000 UTC" firstStartedPulling="2025-11-23 07:13:00.820663591 +0000 UTC m=+1396.334054894" lastFinishedPulling="2025-11-23 07:13:03.477800684 +0000 UTC m=+1398.991192007" observedRunningTime="2025-11-23 07:13:03.911374335 +0000 UTC m=+1399.424765638" watchObservedRunningTime="2025-11-23 07:13:03.916295604 +0000 UTC m=+1399.429686907" Nov 23 07:13:04 crc kubenswrapper[4906]: I1123 07:13:04.581369 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8r2d9" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="registry-server" probeResult="failure" output=< Nov 23 07:13:04 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 07:13:04 crc kubenswrapper[4906]: > Nov 23 07:13:05 crc kubenswrapper[4906]: I1123 07:13:05.597292 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:13:05 crc kubenswrapper[4906]: I1123 07:13:05.597886 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:13:06 crc kubenswrapper[4906]: I1123 07:13:06.672984 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 07:13:08 crc kubenswrapper[4906]: I1123 07:13:08.181182 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:13:08 crc kubenswrapper[4906]: I1123 07:13:08.181496 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:13:09 crc kubenswrapper[4906]: I1123 07:13:09.195932 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:13:09 crc kubenswrapper[4906]: I1123 07:13:09.195944 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:13:09 crc kubenswrapper[4906]: I1123 07:13:09.485660 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:13:09 crc kubenswrapper[4906]: I1123 07:13:09.485789 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:13:09 crc kubenswrapper[4906]: I1123 07:13:09.570550 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:13:10 crc kubenswrapper[4906]: I1123 07:13:10.044913 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:13:10 crc kubenswrapper[4906]: I1123 07:13:10.119286 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8pttz"] Nov 23 07:13:10 crc kubenswrapper[4906]: I1123 07:13:10.597848 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:13:10 crc kubenswrapper[4906]: I1123 07:13:10.598025 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:13:11 crc kubenswrapper[4906]: I1123 07:13:11.613943 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:13:11 crc kubenswrapper[4906]: I1123 07:13:11.614077 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:13:11 crc kubenswrapper[4906]: I1123 07:13:11.674108 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 07:13:11 crc kubenswrapper[4906]: I1123 07:13:11.710598 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 07:13:11 crc kubenswrapper[4906]: I1123 07:13:11.977233 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8pttz" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="registry-server" containerID="cri-o://46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321" gracePeriod=2 Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.017783 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.489381 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.613195 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-utilities\") pod \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.613410 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fprc\" (UniqueName: \"kubernetes.io/projected/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-kube-api-access-9fprc\") pod \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.613454 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-catalog-content\") pod \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\" (UID: \"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd\") " Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.614618 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-utilities" (OuterVolumeSpecName: "utilities") pod "adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" (UID: "adeef82e-07c0-45b4-a3c7-0ea6fcd790dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.615342 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.622342 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-kube-api-access-9fprc" (OuterVolumeSpecName: "kube-api-access-9fprc") pod "adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" (UID: "adeef82e-07c0-45b4-a3c7-0ea6fcd790dd"). InnerVolumeSpecName "kube-api-access-9fprc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.677108 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" (UID: "adeef82e-07c0-45b4-a3c7-0ea6fcd790dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.718015 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fprc\" (UniqueName: \"kubernetes.io/projected/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-kube-api-access-9fprc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.718055 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.991486 4906 generic.go:334] "Generic (PLEG): container finished" podID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerID="46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321" exitCode=0 Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.991546 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pttz" event={"ID":"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd","Type":"ContainerDied","Data":"46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321"} Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.991651 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pttz" Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.992145 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pttz" event={"ID":"adeef82e-07c0-45b4-a3c7-0ea6fcd790dd","Type":"ContainerDied","Data":"a5da67ebcb43f72804afba14957ff2601d59462a0739ad7d044da4f610882f0c"} Nov 23 07:13:12 crc kubenswrapper[4906]: I1123 07:13:12.992187 4906 scope.go:117] "RemoveContainer" containerID="46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.090988 4906 scope.go:117] "RemoveContainer" containerID="0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.097790 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8pttz"] Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.107650 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8pttz"] Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.128127 4906 scope.go:117] "RemoveContainer" containerID="28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.184148 4906 scope.go:117] "RemoveContainer" containerID="46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321" Nov 23 07:13:13 crc kubenswrapper[4906]: E1123 07:13:13.185267 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321\": container with ID starting with 46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321 not found: ID does not exist" containerID="46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.185312 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321"} err="failed to get container status \"46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321\": rpc error: code = NotFound desc = could not find container \"46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321\": container with ID starting with 46e3dfbf22678453bc41e7cb671e6a58f346a8274ca617e6a7ec04316bc15321 not found: ID does not exist" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.185343 4906 scope.go:117] "RemoveContainer" containerID="0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6" Nov 23 07:13:13 crc kubenswrapper[4906]: E1123 07:13:13.185917 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6\": container with ID starting with 0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6 not found: ID does not exist" containerID="0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.185953 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6"} err="failed to get container status \"0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6\": rpc error: code = NotFound desc = could not find container \"0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6\": container with ID starting with 0f5b6aa87560ebb37b7f1480a03d3af955da9cc728d4b6584db534eaff8cb7b6 not found: ID does not exist" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.185982 4906 scope.go:117] "RemoveContainer" containerID="28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc" Nov 23 07:13:13 crc kubenswrapper[4906]: E1123 07:13:13.186549 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc\": container with ID starting with 28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc not found: ID does not exist" containerID="28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.186607 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc"} err="failed to get container status \"28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc\": rpc error: code = NotFound desc = could not find container \"28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc\": container with ID starting with 28d19774f07efc9f65956ed1e82713312f83073caf2f28fe6278a9b777a5dcbc not found: ID does not exist" Nov 23 07:13:13 crc kubenswrapper[4906]: I1123 07:13:13.370468 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" path="/var/lib/kubelet/pods/adeef82e-07c0-45b4-a3c7-0ea6fcd790dd/volumes" Nov 23 07:13:14 crc kubenswrapper[4906]: I1123 07:13:14.531896 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8r2d9" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="registry-server" probeResult="failure" output=< Nov 23 07:13:14 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 07:13:14 crc kubenswrapper[4906]: > Nov 23 07:13:18 crc kubenswrapper[4906]: I1123 07:13:18.189986 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:13:18 crc kubenswrapper[4906]: I1123 07:13:18.191418 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:13:18 crc kubenswrapper[4906]: I1123 07:13:18.191469 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:13:18 crc kubenswrapper[4906]: I1123 07:13:18.202064 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:13:18 crc kubenswrapper[4906]: I1123 07:13:18.993043 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 07:13:19 crc kubenswrapper[4906]: I1123 07:13:19.058832 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:13:19 crc kubenswrapper[4906]: I1123 07:13:19.067813 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.292168 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lj9dt"] Nov 23 07:13:20 crc kubenswrapper[4906]: E1123 07:13:20.292581 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="registry-server" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.292597 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="registry-server" Nov 23 07:13:20 crc kubenswrapper[4906]: E1123 07:13:20.292635 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="extract-content" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.292641 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="extract-content" Nov 23 07:13:20 crc kubenswrapper[4906]: E1123 07:13:20.292660 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="extract-utilities" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.292667 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="extract-utilities" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.292870 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="adeef82e-07c0-45b4-a3c7-0ea6fcd790dd" containerName="registry-server" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.294981 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.310395 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj9dt"] Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.438272 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-catalog-content\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.438611 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcf59\" (UniqueName: \"kubernetes.io/projected/078c4636-e577-4b77-9bc9-4d878dee5313-kube-api-access-jcf59\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.438672 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-utilities\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.540419 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-catalog-content\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.540746 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcf59\" (UniqueName: \"kubernetes.io/projected/078c4636-e577-4b77-9bc9-4d878dee5313-kube-api-access-jcf59\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.540795 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-utilities\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.542444 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-utilities\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.542600 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-catalog-content\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.566474 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcf59\" (UniqueName: \"kubernetes.io/projected/078c4636-e577-4b77-9bc9-4d878dee5313-kube-api-access-jcf59\") pod \"redhat-marketplace-lj9dt\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.603768 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.605231 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.620742 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.628515 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.945438 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:13:20 crc kubenswrapper[4906]: I1123 07:13:20.945790 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:13:21 crc kubenswrapper[4906]: I1123 07:13:21.083244 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:13:21 crc kubenswrapper[4906]: I1123 07:13:21.197871 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj9dt"] Nov 23 07:13:21 crc kubenswrapper[4906]: W1123 07:13:21.213236 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod078c4636_e577_4b77_9bc9_4d878dee5313.slice/crio-cc4600875e71b45ca250721dff596e2034c62f7feed7c48eeb9877abe069823b WatchSource:0}: Error finding container cc4600875e71b45ca250721dff596e2034c62f7feed7c48eeb9877abe069823b: Status 404 returned error can't find the container with id cc4600875e71b45ca250721dff596e2034c62f7feed7c48eeb9877abe069823b Nov 23 07:13:22 crc kubenswrapper[4906]: I1123 07:13:22.093324 4906 generic.go:334] "Generic (PLEG): container finished" podID="078c4636-e577-4b77-9bc9-4d878dee5313" containerID="b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46" exitCode=0 Nov 23 07:13:22 crc kubenswrapper[4906]: I1123 07:13:22.093522 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj9dt" event={"ID":"078c4636-e577-4b77-9bc9-4d878dee5313","Type":"ContainerDied","Data":"b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46"} Nov 23 07:13:22 crc kubenswrapper[4906]: I1123 07:13:22.094235 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj9dt" event={"ID":"078c4636-e577-4b77-9bc9-4d878dee5313","Type":"ContainerStarted","Data":"cc4600875e71b45ca250721dff596e2034c62f7feed7c48eeb9877abe069823b"} Nov 23 07:13:23 crc kubenswrapper[4906]: I1123 07:13:23.110330 4906 generic.go:334] "Generic (PLEG): container finished" podID="078c4636-e577-4b77-9bc9-4d878dee5313" containerID="f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969" exitCode=0 Nov 23 07:13:23 crc kubenswrapper[4906]: I1123 07:13:23.110445 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj9dt" event={"ID":"078c4636-e577-4b77-9bc9-4d878dee5313","Type":"ContainerDied","Data":"f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969"} Nov 23 07:13:23 crc kubenswrapper[4906]: I1123 07:13:23.554314 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:13:23 crc kubenswrapper[4906]: I1123 07:13:23.623223 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:13:24 crc kubenswrapper[4906]: I1123 07:13:24.135438 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj9dt" event={"ID":"078c4636-e577-4b77-9bc9-4d878dee5313","Type":"ContainerStarted","Data":"2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151"} Nov 23 07:13:24 crc kubenswrapper[4906]: I1123 07:13:24.165716 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lj9dt" podStartSLOduration=2.744123432 podStartE2EDuration="4.16565737s" podCreationTimestamp="2025-11-23 07:13:20 +0000 UTC" firstStartedPulling="2025-11-23 07:13:22.097116429 +0000 UTC m=+1417.610507732" lastFinishedPulling="2025-11-23 07:13:23.518650367 +0000 UTC m=+1419.032041670" observedRunningTime="2025-11-23 07:13:24.161787778 +0000 UTC m=+1419.675179091" watchObservedRunningTime="2025-11-23 07:13:24.16565737 +0000 UTC m=+1419.679048683" Nov 23 07:13:25 crc kubenswrapper[4906]: I1123 07:13:25.459925 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8r2d9"] Nov 23 07:13:25 crc kubenswrapper[4906]: I1123 07:13:25.460714 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8r2d9" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="registry-server" containerID="cri-o://da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8" gracePeriod=2 Nov 23 07:13:25 crc kubenswrapper[4906]: I1123 07:13:25.964127 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.088313 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-catalog-content\") pod \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.088461 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skg5c\" (UniqueName: \"kubernetes.io/projected/c2432dda-e7b8-4e36-823c-fa1a3075b57b-kube-api-access-skg5c\") pod \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.089309 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-utilities\") pod \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\" (UID: \"c2432dda-e7b8-4e36-823c-fa1a3075b57b\") " Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.090999 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-utilities" (OuterVolumeSpecName: "utilities") pod "c2432dda-e7b8-4e36-823c-fa1a3075b57b" (UID: "c2432dda-e7b8-4e36-823c-fa1a3075b57b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.093010 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.095917 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2432dda-e7b8-4e36-823c-fa1a3075b57b-kube-api-access-skg5c" (OuterVolumeSpecName: "kube-api-access-skg5c") pod "c2432dda-e7b8-4e36-823c-fa1a3075b57b" (UID: "c2432dda-e7b8-4e36-823c-fa1a3075b57b"). InnerVolumeSpecName "kube-api-access-skg5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.166825 4906 generic.go:334] "Generic (PLEG): container finished" podID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerID="da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8" exitCode=0 Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.166915 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r2d9" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.166920 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r2d9" event={"ID":"c2432dda-e7b8-4e36-823c-fa1a3075b57b","Type":"ContainerDied","Data":"da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8"} Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.167133 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r2d9" event={"ID":"c2432dda-e7b8-4e36-823c-fa1a3075b57b","Type":"ContainerDied","Data":"b85b4dd29efddcd027497a37381c56b53e43347c81ad7291ba1209607e17f70d"} Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.167223 4906 scope.go:117] "RemoveContainer" containerID="da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.192948 4906 scope.go:117] "RemoveContainer" containerID="f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.197972 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skg5c\" (UniqueName: \"kubernetes.io/projected/c2432dda-e7b8-4e36-823c-fa1a3075b57b-kube-api-access-skg5c\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.204658 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2432dda-e7b8-4e36-823c-fa1a3075b57b" (UID: "c2432dda-e7b8-4e36-823c-fa1a3075b57b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.221874 4906 scope.go:117] "RemoveContainer" containerID="14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.281123 4906 scope.go:117] "RemoveContainer" containerID="da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8" Nov 23 07:13:26 crc kubenswrapper[4906]: E1123 07:13:26.281944 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8\": container with ID starting with da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8 not found: ID does not exist" containerID="da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.282018 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8"} err="failed to get container status \"da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8\": rpc error: code = NotFound desc = could not find container \"da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8\": container with ID starting with da6cca2fbc5ce95e5ec3d488040c27902006bb887a40e1c2f32e13668fae5fb8 not found: ID does not exist" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.282061 4906 scope.go:117] "RemoveContainer" containerID="f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7" Nov 23 07:13:26 crc kubenswrapper[4906]: E1123 07:13:26.282513 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7\": container with ID starting with f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7 not found: ID does not exist" containerID="f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.282618 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7"} err="failed to get container status \"f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7\": rpc error: code = NotFound desc = could not find container \"f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7\": container with ID starting with f0b70335ccb8e2d40c8fc5770ac109ec802e34d196bd0d6d4bd335628927f5e7 not found: ID does not exist" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.282740 4906 scope.go:117] "RemoveContainer" containerID="14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4" Nov 23 07:13:26 crc kubenswrapper[4906]: E1123 07:13:26.283114 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4\": container with ID starting with 14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4 not found: ID does not exist" containerID="14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.283150 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4"} err="failed to get container status \"14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4\": rpc error: code = NotFound desc = could not find container \"14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4\": container with ID starting with 14296a9ded18d283efede49468634c3e53855251c5a4fda5c460fd50a569dec4 not found: ID does not exist" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.307632 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2432dda-e7b8-4e36-823c-fa1a3075b57b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.515495 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8r2d9"] Nov 23 07:13:26 crc kubenswrapper[4906]: I1123 07:13:26.523098 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8r2d9"] Nov 23 07:13:27 crc kubenswrapper[4906]: I1123 07:13:27.378436 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" path="/var/lib/kubelet/pods/c2432dda-e7b8-4e36-823c-fa1a3075b57b/volumes" Nov 23 07:13:30 crc kubenswrapper[4906]: I1123 07:13:30.628853 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:30 crc kubenswrapper[4906]: I1123 07:13:30.629283 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:30 crc kubenswrapper[4906]: I1123 07:13:30.699214 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:31 crc kubenswrapper[4906]: I1123 07:13:31.300249 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:31 crc kubenswrapper[4906]: I1123 07:13:31.857899 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj9dt"] Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.275479 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lj9dt" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="registry-server" containerID="cri-o://2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151" gracePeriod=2 Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.774108 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.888270 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-catalog-content\") pod \"078c4636-e577-4b77-9bc9-4d878dee5313\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.888362 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcf59\" (UniqueName: \"kubernetes.io/projected/078c4636-e577-4b77-9bc9-4d878dee5313-kube-api-access-jcf59\") pod \"078c4636-e577-4b77-9bc9-4d878dee5313\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.888493 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-utilities\") pod \"078c4636-e577-4b77-9bc9-4d878dee5313\" (UID: \"078c4636-e577-4b77-9bc9-4d878dee5313\") " Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.889812 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-utilities" (OuterVolumeSpecName: "utilities") pod "078c4636-e577-4b77-9bc9-4d878dee5313" (UID: "078c4636-e577-4b77-9bc9-4d878dee5313"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.899338 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/078c4636-e577-4b77-9bc9-4d878dee5313-kube-api-access-jcf59" (OuterVolumeSpecName: "kube-api-access-jcf59") pod "078c4636-e577-4b77-9bc9-4d878dee5313" (UID: "078c4636-e577-4b77-9bc9-4d878dee5313"). InnerVolumeSpecName "kube-api-access-jcf59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.905502 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "078c4636-e577-4b77-9bc9-4d878dee5313" (UID: "078c4636-e577-4b77-9bc9-4d878dee5313"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.990700 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.990728 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078c4636-e577-4b77-9bc9-4d878dee5313-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:33 crc kubenswrapper[4906]: I1123 07:13:33.990738 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcf59\" (UniqueName: \"kubernetes.io/projected/078c4636-e577-4b77-9bc9-4d878dee5313-kube-api-access-jcf59\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.323156 4906 generic.go:334] "Generic (PLEG): container finished" podID="078c4636-e577-4b77-9bc9-4d878dee5313" containerID="2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151" exitCode=0 Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.323268 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lj9dt" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.323296 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj9dt" event={"ID":"078c4636-e577-4b77-9bc9-4d878dee5313","Type":"ContainerDied","Data":"2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151"} Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.323701 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lj9dt" event={"ID":"078c4636-e577-4b77-9bc9-4d878dee5313","Type":"ContainerDied","Data":"cc4600875e71b45ca250721dff596e2034c62f7feed7c48eeb9877abe069823b"} Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.323790 4906 scope.go:117] "RemoveContainer" containerID="2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.362967 4906 scope.go:117] "RemoveContainer" containerID="f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.368339 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj9dt"] Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.375915 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lj9dt"] Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.404837 4906 scope.go:117] "RemoveContainer" containerID="b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.448259 4906 scope.go:117] "RemoveContainer" containerID="2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151" Nov 23 07:13:34 crc kubenswrapper[4906]: E1123 07:13:34.449455 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151\": container with ID starting with 2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151 not found: ID does not exist" containerID="2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.449829 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151"} err="failed to get container status \"2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151\": rpc error: code = NotFound desc = could not find container \"2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151\": container with ID starting with 2ac3a5ac484a4a160898cf184965ab0d7d5153daff2417a9ef0ebf9fc99ab151 not found: ID does not exist" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.450122 4906 scope.go:117] "RemoveContainer" containerID="f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969" Nov 23 07:13:34 crc kubenswrapper[4906]: E1123 07:13:34.451294 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969\": container with ID starting with f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969 not found: ID does not exist" containerID="f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.451343 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969"} err="failed to get container status \"f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969\": rpc error: code = NotFound desc = could not find container \"f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969\": container with ID starting with f702cbd33a61e8fd76ebca3fdd32aa3aa820b67ebf81953cb4275a1bbfa59969 not found: ID does not exist" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.451376 4906 scope.go:117] "RemoveContainer" containerID="b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46" Nov 23 07:13:34 crc kubenswrapper[4906]: E1123 07:13:34.451969 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46\": container with ID starting with b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46 not found: ID does not exist" containerID="b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46" Nov 23 07:13:34 crc kubenswrapper[4906]: I1123 07:13:34.452198 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46"} err="failed to get container status \"b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46\": rpc error: code = NotFound desc = could not find container \"b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46\": container with ID starting with b6b3e325cb884f146294bb01516ec1995393de827f357fc1d46206cccc5a5a46 not found: ID does not exist" Nov 23 07:13:35 crc kubenswrapper[4906]: I1123 07:13:35.373281 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" path="/var/lib/kubelet/pods/078c4636-e577-4b77-9bc9-4d878dee5313/volumes" Nov 23 07:13:40 crc kubenswrapper[4906]: I1123 07:13:40.595991 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 23 07:13:40 crc kubenswrapper[4906]: I1123 07:13:40.596946 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" containerName="openstackclient" containerID="cri-o://6bc391c38da2e27698e0bb15d809a3a29934990091bf87fa84600dc4d231301b" gracePeriod=2 Nov 23 07:13:40 crc kubenswrapper[4906]: I1123 07:13:40.625094 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 23 07:13:40 crc kubenswrapper[4906]: I1123 07:13:40.902156 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-ftw6h"] Nov 23 07:13:40 crc kubenswrapper[4906]: I1123 07:13:40.951964 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l9d4x"] Nov 23 07:13:40 crc kubenswrapper[4906]: I1123 07:13:40.999785 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.031808 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicana450-account-delete-l2bs5"] Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.032471 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="extract-utilities" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032486 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="extract-utilities" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.032503 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" containerName="openstackclient" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032509 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" containerName="openstackclient" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.032529 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="registry-server" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032535 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="registry-server" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.032551 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="registry-server" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032557 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="registry-server" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.032574 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="extract-content" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032579 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="extract-content" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.032587 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="extract-content" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032592 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="extract-content" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.032615 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="extract-utilities" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032620 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="extract-utilities" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032824 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" containerName="openstackclient" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032842 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="078c4636-e577-4b77-9bc9-4d878dee5313" containerName="registry-server" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.032871 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2432dda-e7b8-4e36-823c-fa1a3075b57b" containerName="registry-server" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.033671 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.040790 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicana450-account-delete-l2bs5"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.072269 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-njmjc"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.072571 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-njmjc" podUID="3acf9ff7-0012-4e1a-b940-037bebcc4c88" containerName="openstack-network-exporter" containerID="cri-o://28d9044f3cc3149e597f205efe1d60660461a96061a5b3c9310711ba4008ce47" gracePeriod=30 Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.075605 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.075673 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data podName:4d677a57-1acf-4627-aa66-f4bee96a2b51 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:41.575655057 +0000 UTC m=+1437.089046350 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data") pod "rabbitmq-cell1-server-0" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51") : configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.167505 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementb64f-account-delete-576n6"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.169181 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.205670 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts\") pod \"barbicana450-account-delete-l2bs5\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.205872 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsw9k\" (UniqueName: \"kubernetes.io/projected/64c1c989-3443-4b80-9786-56b68660655c-kube-api-access-rsw9k\") pod \"barbicana450-account-delete-l2bs5\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.215929 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementb64f-account-delete-576n6"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.277799 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9l4vd"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.308317 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6093f84a-c555-4fc4-a207-f54e5ddf2a85-operator-scripts\") pod \"placementb64f-account-delete-576n6\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.308410 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts\") pod \"barbicana450-account-delete-l2bs5\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.308469 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmvjr\" (UniqueName: \"kubernetes.io/projected/6093f84a-c555-4fc4-a207-f54e5ddf2a85-kube-api-access-tmvjr\") pod \"placementb64f-account-delete-576n6\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.308528 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsw9k\" (UniqueName: \"kubernetes.io/projected/64c1c989-3443-4b80-9786-56b68660655c-kube-api-access-rsw9k\") pod \"barbicana450-account-delete-l2bs5\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.309185 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9l4vd"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.309690 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts\") pod \"barbicana450-account-delete-l2bs5\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.384598 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5584696b-3782-4db6-997e-9f691f70d05a" path="/var/lib/kubelet/pods/5584696b-3782-4db6-997e-9f691f70d05a/volumes" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.385722 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.394262 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="ovn-northd" containerID="cri-o://e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" gracePeriod=30 Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.394893 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="openstack-network-exporter" containerID="cri-o://631998df3fcc19a80bacb62ef5030d483398f5a45ac22c6b6964329c20351f4a" gracePeriod=30 Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.396605 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsw9k\" (UniqueName: \"kubernetes.io/projected/64c1c989-3443-4b80-9786-56b68660655c-kube-api-access-rsw9k\") pod \"barbicana450-account-delete-l2bs5\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.405840 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder46ad-account-delete-rpn4p"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.410553 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.412393 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6093f84a-c555-4fc4-a207-f54e5ddf2a85-operator-scripts\") pod \"placementb64f-account-delete-576n6\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.412601 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmvjr\" (UniqueName: \"kubernetes.io/projected/6093f84a-c555-4fc4-a207-f54e5ddf2a85-kube-api-access-tmvjr\") pod \"placementb64f-account-delete-576n6\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.413830 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6093f84a-c555-4fc4-a207-f54e5ddf2a85-operator-scripts\") pod \"placementb64f-account-delete-576n6\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.452327 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder46ad-account-delete-rpn4p"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.466503 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmvjr\" (UniqueName: \"kubernetes.io/projected/6093f84a-c555-4fc4-a207-f54e5ddf2a85-kube-api-access-tmvjr\") pod \"placementb64f-account-delete-576n6\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.494927 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.514091 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvxnr\" (UniqueName: \"kubernetes.io/projected/4df634a0-630f-46a5-b304-86211157136a-kube-api-access-hvxnr\") pod \"cinder46ad-account-delete-rpn4p\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.514509 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4df634a0-630f-46a5-b304-86211157136a-operator-scripts\") pod \"cinder46ad-account-delete-rpn4p\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.585428 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.620509 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4df634a0-630f-46a5-b304-86211157136a-operator-scripts\") pod \"cinder46ad-account-delete-rpn4p\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.622536 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvxnr\" (UniqueName: \"kubernetes.io/projected/4df634a0-630f-46a5-b304-86211157136a-kube-api-access-hvxnr\") pod \"cinder46ad-account-delete-rpn4p\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.624134 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.624196 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data podName:9a0222ea-7767-4d08-a6ec-6659a33f9df2 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:42.12417731 +0000 UTC m=+1437.637568613 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data") pod "rabbitmq-server-0" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2") : configmap "rabbitmq-config-data" not found Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.625095 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4df634a0-630f-46a5-b304-86211157136a-operator-scripts\") pod \"cinder46ad-account-delete-rpn4p\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.626880 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.626961 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data podName:4d677a57-1acf-4627-aa66-f4bee96a2b51 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:42.626930752 +0000 UTC m=+1438.140322055 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data") pod "rabbitmq-cell1-server-0" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51") : configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.674835 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.675925 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6wvm5"] Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.719509 4906 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-l9d4x" message=< Nov 23 07:13:41 crc kubenswrapper[4906]: Exiting ovn-controller (1) [ OK ] Nov 23 07:13:41 crc kubenswrapper[4906]: > Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.719557 4906 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-l9d4x" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" containerID="cri-o://5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.719608 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-l9d4x" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" containerID="cri-o://5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e" gracePeriod=30 Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.725700 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvxnr\" (UniqueName: \"kubernetes.io/projected/4df634a0-630f-46a5-b304-86211157136a-kube-api-access-hvxnr\") pod \"cinder46ad-account-delete-rpn4p\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.753242 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e is running failed: container process not found" containerID="5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.759904 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e is running failed: container process not found" containerID="5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.771773 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6wvm5"] Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.785124 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e is running failed: container process not found" containerID="5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 23 07:13:41 crc kubenswrapper[4906]: E1123 07:13:41.788179 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-l9d4x" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.829575 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.851733 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-6p4bd"] Nov 23 07:13:41 crc kubenswrapper[4906]: I1123 07:13:41.954557 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-6p4bd"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.010692 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glanceaaf0-account-delete-j59k7"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.012227 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.051789 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glanceaaf0-account-delete-j59k7"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.114765 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e0ab9b5-09c4-49b9-a112-76c18f06524d-operator-scripts\") pod \"glanceaaf0-account-delete-j59k7\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.115143 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p9gc\" (UniqueName: \"kubernetes.io/projected/2e0ab9b5-09c4-49b9-a112-76c18f06524d-kube-api-access-9p9gc\") pod \"glanceaaf0-account-delete-j59k7\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.184825 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron9f60-account-delete-vq457"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.186626 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.235439 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts\") pod \"neutron9f60-account-delete-vq457\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.235693 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p9gc\" (UniqueName: \"kubernetes.io/projected/2e0ab9b5-09c4-49b9-a112-76c18f06524d-kube-api-access-9p9gc\") pod \"glanceaaf0-account-delete-j59k7\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.235731 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqcfw\" (UniqueName: \"kubernetes.io/projected/00991da6-745e-4757-9d89-a20c5b97e38f-kube-api-access-rqcfw\") pod \"neutron9f60-account-delete-vq457\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.235988 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e0ab9b5-09c4-49b9-a112-76c18f06524d-operator-scripts\") pod \"glanceaaf0-account-delete-j59k7\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.236933 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e0ab9b5-09c4-49b9-a112-76c18f06524d-operator-scripts\") pod \"glanceaaf0-account-delete-j59k7\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: E1123 07:13:42.237029 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 23 07:13:42 crc kubenswrapper[4906]: E1123 07:13:42.237079 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data podName:9a0222ea-7767-4d08-a6ec-6659a33f9df2 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:43.23706459 +0000 UTC m=+1438.750455893 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data") pod "rabbitmq-server-0" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2") : configmap "rabbitmq-config-data" not found Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.260506 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron9f60-account-delete-vq457"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.297229 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-d2j47"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.333570 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.334528 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="openstack-network-exporter" containerID="cri-o://a091f675900348c0b86ab0b4409f0e26e962032b4b02dd301de38636bc1e7e50" gracePeriod=300 Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.342579 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts\") pod \"neutron9f60-account-delete-vq457\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.342730 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqcfw\" (UniqueName: \"kubernetes.io/projected/00991da6-745e-4757-9d89-a20c5b97e38f-kube-api-access-rqcfw\") pod \"neutron9f60-account-delete-vq457\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.344067 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts\") pod \"neutron9f60-account-delete-vq457\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.371032 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-d2j47"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.446111 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-944h6"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.455423 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerName="dnsmasq-dns" containerID="cri-o://9310cf8e52799cd9ffd94d046c01ec36de9b4fb21b64eff04510a5c86bb2a2cd" gracePeriod=10 Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.507374 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell082ff-account-delete-s8jc5"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.519299 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.545131 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell082ff-account-delete-s8jc5"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.552436 4906 generic.go:334] "Generic (PLEG): container finished" podID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerID="631998df3fcc19a80bacb62ef5030d483398f5a45ac22c6b6964329c20351f4a" exitCode=2 Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.552532 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63999fb5-88e7-4df1-8084-267d0e37ac4c","Type":"ContainerDied","Data":"631998df3fcc19a80bacb62ef5030d483398f5a45ac22c6b6964329c20351f4a"} Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.574711 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqcfw\" (UniqueName: \"kubernetes.io/projected/00991da6-745e-4757-9d89-a20c5b97e38f-kube-api-access-rqcfw\") pod \"neutron9f60-account-delete-vq457\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.611206 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p9gc\" (UniqueName: \"kubernetes.io/projected/2e0ab9b5-09c4-49b9-a112-76c18f06524d-kube-api-access-9p9gc\") pod \"glanceaaf0-account-delete-j59k7\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.651282 4906 generic.go:334] "Generic (PLEG): container finished" podID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerID="5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e" exitCode=0 Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.651430 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x" event={"ID":"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd","Type":"ContainerDied","Data":"5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e"} Nov 23 07:13:42 crc kubenswrapper[4906]: E1123 07:13:42.653608 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:42 crc kubenswrapper[4906]: E1123 07:13:42.654057 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data podName:4d677a57-1acf-4627-aa66-f4bee96a2b51 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:44.654037556 +0000 UTC m=+1440.167428859 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data") pod "rabbitmq-cell1-server-0" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51") : configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.658725 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.677232 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.678116 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="openstack-network-exporter" containerID="cri-o://a901c2249d99c2e7c335898965028b8ed5f7a9b34574ccbb6be133b51490d695" gracePeriod=300 Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.692365 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-njmjc_3acf9ff7-0012-4e1a-b940-037bebcc4c88/openstack-network-exporter/0.log" Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.692417 4906 generic.go:334] "Generic (PLEG): container finished" podID="3acf9ff7-0012-4e1a-b940-037bebcc4c88" containerID="28d9044f3cc3149e597f205efe1d60660461a96061a5b3c9310711ba4008ce47" exitCode=2 Nov 23 07:13:42 crc kubenswrapper[4906]: I1123 07:13:42.692451 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-njmjc" event={"ID":"3acf9ff7-0012-4e1a-b940-037bebcc4c88","Type":"ContainerDied","Data":"28d9044f3cc3149e597f205efe1d60660461a96061a5b3c9310711ba4008ce47"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.760621 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts\") pod \"novacell082ff-account-delete-s8jc5\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.760824 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxrdn\" (UniqueName: \"kubernetes.io/projected/01d907ea-d712-40af-b18a-5d55287fbaeb-kube-api-access-gxrdn\") pod \"novacell082ff-account-delete-s8jc5\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.834762 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi944e-account-delete-gw8gk"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.839487 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.846790 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ffb7cbc86-lz7bc"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.847207 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-ffb7cbc86-lz7bc" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-log" containerID="cri-o://a3add9b8cbef96dced17c8263087b9f52097f904de7cf1efc850f3e41c400acf" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.847457 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-ffb7cbc86-lz7bc" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-api" containerID="cri-o://e5fc6b77a408f81d5b91fd45ba27278acb2c5d247b602dbc2475870c5ba45b81" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.865334 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-ltbgj"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.865609 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxrdn\" (UniqueName: \"kubernetes.io/projected/01d907ea-d712-40af-b18a-5d55287fbaeb-kube-api-access-gxrdn\") pod \"novacell082ff-account-delete-s8jc5\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.865739 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts\") pod \"novacell082ff-account-delete-s8jc5\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.867087 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts\") pod \"novacell082ff-account-delete-s8jc5\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.895740 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-tcrg7"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.915352 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxrdn\" (UniqueName: \"kubernetes.io/projected/01d907ea-d712-40af-b18a-5d55287fbaeb-kube-api-access-gxrdn\") pod \"novacell082ff-account-delete-s8jc5\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.936076 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-ltbgj"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:42.966966 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-tcrg7"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.001722 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkbkq\" (UniqueName: \"kubernetes.io/projected/0ab42792-22e3-48ee-8028-3d181679190c-kube-api-access-vkbkq\") pod \"novaapi944e-account-delete-gw8gk\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.001919 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab42792-22e3-48ee-8028-3d181679190c-operator-scripts\") pod \"novaapi944e-account-delete-gw8gk\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.009222 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi944e-account-delete-gw8gk"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.033281 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.039266 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-server" containerID="cri-o://14e047c6504017d7ca6ee4d097e4dc0c1a3d8b79305ee2e36580fe0c8238c99b" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.039773 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-server" containerID="cri-o://0c11306e0c7841c07b12efafe09a24ad91a89cee793dba18a9f2876f0d47db2f" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.039963 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-updater" containerID="cri-o://627197269c9d973c5e754357b5bea2cd6202f4ed8b342e5be634b4ffe2fc24f4" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.039983 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="rsync" containerID="cri-o://c1437bfa849a403a91c6260f5cd5aec30c660352ad87626d59e56d5664d49e22" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.039963 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="swift-recon-cron" containerID="cri-o://0f3631e076f67c96972eae5e4520d40c2b4e9bafcc51ca7093e8a282b2d1ab2d" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040046 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-expirer" containerID="cri-o://18a0cbb3bc39086918add992d8b7e36417c2a512ce34ad58af2473769333d276" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040066 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-auditor" containerID="cri-o://59f81af64ef288c720feec11a22a60062d378a09a432fe305317b0c5b6c0c0a3" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040094 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-updater" containerID="cri-o://7310698561fe415b603d4b23e1ca0b97cb10e0a5ab1b9e756910372da0037e1f" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040113 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-replicator" containerID="cri-o://d7765fe2025af755e1f8644c577f06f2002c8f768fc520d7b9ae7dfe318bd069" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040135 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-auditor" containerID="cri-o://52c1da52f5e6b3969985d8979a7d057a2abc04e85e14b228a48cb576460bae01" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040180 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-replicator" containerID="cri-o://28356b85be5f4f2918de3e09fcd6083deb2b49f2e911a40c85312e0eebf206a8" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040206 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-auditor" containerID="cri-o://29fc0d262044ca645fb4133e01cbabc9655fe5b59a91b3028e0d08bcd5278256" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040243 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-reaper" containerID="cri-o://11af2df38e7407624c492ea4718898da13ce0775329468c7c775d8b83ef7a458" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.040269 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-replicator" containerID="cri-o://8bed0393fc8307e59d410b24ec754b21fe7d3bd1f3e83cd8d6f028d1a8f06cdd" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.037445 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-server" containerID="cri-o://4f7149c3b0589b8222d668111ff70a5b576f33041ccbfe291e9f62085cc511d3" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.111231 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkbkq\" (UniqueName: \"kubernetes.io/projected/0ab42792-22e3-48ee-8028-3d181679190c-kube-api-access-vkbkq\") pod \"novaapi944e-account-delete-gw8gk\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.111331 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab42792-22e3-48ee-8028-3d181679190c-operator-scripts\") pod \"novaapi944e-account-delete-gw8gk\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.112453 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab42792-22e3-48ee-8028-3d181679190c-operator-scripts\") pod \"novaapi944e-account-delete-gw8gk\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.127420 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.128650 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a7788825-483c-46f2-a975-239254435019" containerName="cinder-scheduler" containerID="cri-o://209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.129369 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a7788825-483c-46f2-a975-239254435019" containerName="probe" containerID="cri-o://9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.199163 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.199612 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-log" containerID="cri-o://9c4e48b39c14772f337cfc58ea53a79082fd8c022fe4b10170edefdfb0a64859" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.200159 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-httpd" containerID="cri-o://2fd16719fcd506b038a0152208792ed142945112e8297ebc99ad99b173d43f11" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.218380 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="ovsdbserver-sb" containerID="cri-o://c10b96182b460a717b72550d27ff7aae433f2b3f9e4bf71a704dc8b41e9653fb" gracePeriod=300 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.235775 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkbkq\" (UniqueName: \"kubernetes.io/projected/0ab42792-22e3-48ee-8028-3d181679190c-kube-api-access-vkbkq\") pod \"novaapi944e-account-delete-gw8gk\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.259470 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-h4lvc"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.267571 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="ovsdbserver-nb" containerID="cri-o://2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0" gracePeriod=300 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.286647 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.287080 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api-log" containerID="cri-o://df25a67d5a9f5925815eee2060a4fdcc509bfcf08b8d1d1a7f6827a40492c808" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.287287 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api" containerID="cri-o://e82535f708e61c7ee739e897ea48ec09a393ccd402bebffaf681b12cea2dcb54" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.308782 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-h4lvc"] Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.329574 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.329662 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data podName:9a0222ea-7767-4d08-a6ec-6659a33f9df2 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:45.329642128 +0000 UTC m=+1440.843033421 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data") pod "rabbitmq-server-0" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2") : configmap "rabbitmq-config-data" not found Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.353851 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-flw9z"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.376558 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cc86f52-0139-430f-9803-06b714d9fc7e" path="/var/lib/kubelet/pods/0cc86f52-0139-430f-9803-06b714d9fc7e/volumes" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.379828 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32ddb74f-97de-4c04-bdca-efbcf93e5840" path="/var/lib/kubelet/pods/32ddb74f-97de-4c04-bdca-efbcf93e5840/volumes" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.385443 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35bc1bb0-6a22-4beb-bd74-111e73116437" path="/var/lib/kubelet/pods/35bc1bb0-6a22-4beb-bd74-111e73116437/volumes" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.386140 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e15932-ac43-4d8e-ba00-481ad71b5816" path="/var/lib/kubelet/pods/35e15932-ac43-4d8e-ba00-481ad71b5816/volumes" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.386898 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f23c4a7-b2b9-4ef9-876b-5dd95f486333" path="/var/lib/kubelet/pods/8f23c4a7-b2b9-4ef9-876b-5dd95f486333/volumes" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.388292 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2d3d804-9bab-439f-94c6-23e1546aaad7" path="/var/lib/kubelet/pods/d2d3d804-9bab-439f-94c6-23e1546aaad7/volumes" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.394366 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-flw9z"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.405479 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" containerID="cri-o://45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" gracePeriod=28 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.446733 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6f9d47b994-ptdjw"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.447013 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener-log" containerID="cri-o://e5685a80b861544fdb6e63f53e8fc51d7652169e4cb7aa78555e20b2db5bcf62" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.447174 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener" containerID="cri-o://9acde65807d055f979c4c0e765b70fcc9e174d4e93f5057b281b3cd782bbd787" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.507985 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bdb7fc4cb-h962s"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.508927 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bdb7fc4cb-h962s" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api-log" containerID="cri-o://78e2be376a6ac18267e9fa24359c2479618c77eeaf00a6b4b20186ab50f161d6" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.509364 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bdb7fc4cb-h962s" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api" containerID="cri-o://0dd9b19587494f220b9ce7cb80f137d8caf2f6b1707a1f5bae2b3bc3625d12af" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.521121 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5584fdbb5-9b5k8"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.521415 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5584fdbb5-9b5k8" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker-log" containerID="cri-o://9488749cc23de30cd7d4a1998d3ecd858dad70cda1dd749aa047283ebdd924bf" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.521571 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5584fdbb5-9b5k8" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker" containerID="cri-o://04f1ef97bd7ad66e51909908566ff20e01fcea5f37847fa8e0ca7e76df6a1b60" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.544387 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.545564 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-log" containerID="cri-o://85604e28eb73da42df364a73f2a404a0c2cac5adb3cdbd3da52bb12fba8bccc0" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.546534 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-httpd" containerID="cri-o://f4c66a184ddf3fddd4f01f4bf5d9927e08a50644685ee93df83ca8a705617d12" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.553584 4906 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 23 07:13:43 crc kubenswrapper[4906]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 23 07:13:43 crc kubenswrapper[4906]: + source /usr/local/bin/container-scripts/functions Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNBridge=br-int Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNRemote=tcp:localhost:6642 Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNEncapType=geneve Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNAvailabilityZones= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ EnableChassisAsGateway=true Nov 23 07:13:43 crc kubenswrapper[4906]: ++ PhysicalNetworks= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNHostName= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 23 07:13:43 crc kubenswrapper[4906]: ++ ovs_dir=/var/lib/openvswitch Nov 23 07:13:43 crc kubenswrapper[4906]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 23 07:13:43 crc kubenswrapper[4906]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 23 07:13:43 crc kubenswrapper[4906]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + cleanup_ovsdb_server_semaphore Nov 23 07:13:43 crc kubenswrapper[4906]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 23 07:13:43 crc kubenswrapper[4906]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 23 07:13:43 crc kubenswrapper[4906]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-ftw6h" message=< Nov 23 07:13:43 crc kubenswrapper[4906]: Exiting ovsdb-server (5) [ OK ] Nov 23 07:13:43 crc kubenswrapper[4906]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 23 07:13:43 crc kubenswrapper[4906]: + source /usr/local/bin/container-scripts/functions Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNBridge=br-int Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNRemote=tcp:localhost:6642 Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNEncapType=geneve Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNAvailabilityZones= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ EnableChassisAsGateway=true Nov 23 07:13:43 crc kubenswrapper[4906]: ++ PhysicalNetworks= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNHostName= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 23 07:13:43 crc kubenswrapper[4906]: ++ ovs_dir=/var/lib/openvswitch Nov 23 07:13:43 crc kubenswrapper[4906]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 23 07:13:43 crc kubenswrapper[4906]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 23 07:13:43 crc kubenswrapper[4906]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + cleanup_ovsdb_server_semaphore Nov 23 07:13:43 crc kubenswrapper[4906]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 23 07:13:43 crc kubenswrapper[4906]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 23 07:13:43 crc kubenswrapper[4906]: > Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.553634 4906 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 23 07:13:43 crc kubenswrapper[4906]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 23 07:13:43 crc kubenswrapper[4906]: + source /usr/local/bin/container-scripts/functions Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNBridge=br-int Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNRemote=tcp:localhost:6642 Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNEncapType=geneve Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNAvailabilityZones= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ EnableChassisAsGateway=true Nov 23 07:13:43 crc kubenswrapper[4906]: ++ PhysicalNetworks= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ OVNHostName= Nov 23 07:13:43 crc kubenswrapper[4906]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 23 07:13:43 crc kubenswrapper[4906]: ++ ovs_dir=/var/lib/openvswitch Nov 23 07:13:43 crc kubenswrapper[4906]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 23 07:13:43 crc kubenswrapper[4906]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 23 07:13:43 crc kubenswrapper[4906]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + sleep 0.5 Nov 23 07:13:43 crc kubenswrapper[4906]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 23 07:13:43 crc kubenswrapper[4906]: + cleanup_ovsdb_server_semaphore Nov 23 07:13:43 crc kubenswrapper[4906]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 23 07:13:43 crc kubenswrapper[4906]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 23 07:13:43 crc kubenswrapper[4906]: > pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" containerID="cri-o://cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.553693 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" containerID="cri-o://cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" gracePeriod=28 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.576960 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.585529 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f84fc787f-zskfn"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.585829 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f84fc787f-zskfn" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-api" containerID="cri-o://dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.586358 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f84fc787f-zskfn" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-httpd" containerID="cri-o://5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.603909 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-77c4d866c-r9x72"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.604341 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-77c4d866c-r9x72" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-httpd" containerID="cri-o://c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.605078 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-77c4d866c-r9x72" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-server" containerID="cri-o://dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.619042 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.659143 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerName="rabbitmq" containerID="cri-o://d83546ce8058eec4db9f5c5b6441b7ce247b7a874eb9b2018665ca353691af54" gracePeriod=604800 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.673951 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-wbdjb"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.727063 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.731242 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-log" containerID="cri-o://d7113d47184e50909b0d719000d22c051debd5b0dca75d3f957a7612dc3670a1" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.731762 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-api" containerID="cri-o://77ce16e41b99ad023af4aa1620f90f3b6cbe24953a476cead47e6c790b48478b" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.738195 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.743035 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.750635 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 23 07:13:43 crc kubenswrapper[4906]: E1123 07:13:43.750744 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="ovn-northd" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.776752 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.799791 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-wbdjb"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802521 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="c1437bfa849a403a91c6260f5cd5aec30c660352ad87626d59e56d5664d49e22" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802556 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="18a0cbb3bc39086918add992d8b7e36417c2a512ce34ad58af2473769333d276" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802566 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="7310698561fe415b603d4b23e1ca0b97cb10e0a5ab1b9e756910372da0037e1f" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802577 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="52c1da52f5e6b3969985d8979a7d057a2abc04e85e14b228a48cb576460bae01" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802585 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="28356b85be5f4f2918de3e09fcd6083deb2b49f2e911a40c85312e0eebf206a8" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802774 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="14e047c6504017d7ca6ee4d097e4dc0c1a3d8b79305ee2e36580fe0c8238c99b" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802787 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="627197269c9d973c5e754357b5bea2cd6202f4ed8b342e5be634b4ffe2fc24f4" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802794 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="59f81af64ef288c720feec11a22a60062d378a09a432fe305317b0c5b6c0c0a3" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802803 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="d7765fe2025af755e1f8644c577f06f2002c8f768fc520d7b9ae7dfe318bd069" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802811 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="0c11306e0c7841c07b12efafe09a24ad91a89cee793dba18a9f2876f0d47db2f" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802819 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="11af2df38e7407624c492ea4718898da13ce0775329468c7c775d8b83ef7a458" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802828 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="29fc0d262044ca645fb4133e01cbabc9655fe5b59a91b3028e0d08bcd5278256" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802837 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="8bed0393fc8307e59d410b24ec754b21fe7d3bd1f3e83cd8d6f028d1a8f06cdd" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.802845 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="4f7149c3b0589b8222d668111ff70a5b576f33041ccbfe291e9f62085cc511d3" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803162 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"c1437bfa849a403a91c6260f5cd5aec30c660352ad87626d59e56d5664d49e22"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803219 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"18a0cbb3bc39086918add992d8b7e36417c2a512ce34ad58af2473769333d276"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803234 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"7310698561fe415b603d4b23e1ca0b97cb10e0a5ab1b9e756910372da0037e1f"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803244 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"52c1da52f5e6b3969985d8979a7d057a2abc04e85e14b228a48cb576460bae01"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803253 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"28356b85be5f4f2918de3e09fcd6083deb2b49f2e911a40c85312e0eebf206a8"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803262 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"14e047c6504017d7ca6ee4d097e4dc0c1a3d8b79305ee2e36580fe0c8238c99b"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803271 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"627197269c9d973c5e754357b5bea2cd6202f4ed8b342e5be634b4ffe2fc24f4"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803285 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"59f81af64ef288c720feec11a22a60062d378a09a432fe305317b0c5b6c0c0a3"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803294 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"d7765fe2025af755e1f8644c577f06f2002c8f768fc520d7b9ae7dfe318bd069"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803307 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"0c11306e0c7841c07b12efafe09a24ad91a89cee793dba18a9f2876f0d47db2f"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803316 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"11af2df38e7407624c492ea4718898da13ce0775329468c7c775d8b83ef7a458"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803325 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"29fc0d262044ca645fb4133e01cbabc9655fe5b59a91b3028e0d08bcd5278256"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803336 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"8bed0393fc8307e59d410b24ec754b21fe7d3bd1f3e83cd8d6f028d1a8f06cdd"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.803347 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"4f7149c3b0589b8222d668111ff70a5b576f33041ccbfe291e9f62085cc511d3"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.811735 4906 generic.go:334] "Generic (PLEG): container finished" podID="5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" containerID="6bc391c38da2e27698e0bb15d809a3a29934990091bf87fa84600dc4d231301b" exitCode=137 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.812908 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.813272 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-log" containerID="cri-o://cfe0d8e3ca16e113f2016b61cb50f6b5b415f288a07bd395effcd5f6002c7ed6" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.814132 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-metadata" containerID="cri-o://59399848d4fd8c548b0ad9fe61db4277a707a18507241aca88683dc29eebb13a" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.824639 4906 generic.go:334] "Generic (PLEG): container finished" podID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerID="df25a67d5a9f5925815eee2060a4fdcc509bfcf08b8d1d1a7f6827a40492c808" exitCode=143 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.824831 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ee97571-4ae5-4ed0-a32a-67abd5896ae0","Type":"ContainerDied","Data":"df25a67d5a9f5925815eee2060a4fdcc509bfcf08b8d1d1a7f6827a40492c808"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.826530 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.836204 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-3763-account-create-stq74"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.838414 4906 generic.go:334] "Generic (PLEG): container finished" podID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerID="a3add9b8cbef96dced17c8263087b9f52097f904de7cf1efc850f3e41c400acf" exitCode=143 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.838478 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ffb7cbc86-lz7bc" event={"ID":"73149d5c-c16f-4b83-b855-135fc1aab9fe","Type":"ContainerDied","Data":"a3add9b8cbef96dced17c8263087b9f52097f904de7cf1efc850f3e41c400acf"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.842380 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_51dd433b-f1a2-422a-bf5b-643e150c3bca/ovsdbserver-nb/0.log" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.842416 4906 generic.go:334] "Generic (PLEG): container finished" podID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerID="a091f675900348c0b86ab0b4409f0e26e962032b4b02dd301de38636bc1e7e50" exitCode=2 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.842433 4906 generic.go:334] "Generic (PLEG): container finished" podID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerID="2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0" exitCode=143 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.842495 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"51dd433b-f1a2-422a-bf5b-643e150c3bca","Type":"ContainerDied","Data":"a091f675900348c0b86ab0b4409f0e26e962032b4b02dd301de38636bc1e7e50"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.842527 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"51dd433b-f1a2-422a-bf5b-643e150c3bca","Type":"ContainerDied","Data":"2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.852665 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.860466 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-3763-account-create-stq74"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.879059 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.886638 4906 generic.go:334] "Generic (PLEG): container finished" podID="0ce87b9c-add6-4744-8463-0d268867b9de" containerID="e5685a80b861544fdb6e63f53e8fc51d7652169e4cb7aa78555e20b2db5bcf62" exitCode=143 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.886755 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" event={"ID":"0ce87b9c-add6-4744-8463-0d268867b9de","Type":"ContainerDied","Data":"e5685a80b861544fdb6e63f53e8fc51d7652169e4cb7aa78555e20b2db5bcf62"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.890148 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.902793 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.903216 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6cee328a-b746-486d-850b-148ca1e66eea" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://18d1eb931158c01bd8771394295d1f098c58d8673d4374a70905fb56a217df1f" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.904855 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.924132 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.924499 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="da74c07b-2832-4d8e-9508-415fdc3cac71" containerName="nova-scheduler-scheduler" containerID="cri-o://4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.928160 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-njmjc_3acf9ff7-0012-4e1a-b940-037bebcc4c88/openstack-network-exporter/0.log" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.928245 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.944309 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerName="rabbitmq" containerID="cri-o://bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af" gracePeriod=604800 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.950884 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.951134 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="212ee22b-c3e7-4d9c-b952-5837b5982c63" containerName="nova-cell1-conductor-conductor" containerID="cri-o://4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef" gracePeriod=30 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.964513 4906 generic.go:334] "Generic (PLEG): container finished" podID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerID="9310cf8e52799cd9ffd94d046c01ec36de9b4fb21b64eff04510a5c86bb2a2cd" exitCode=0 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.964617 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" event={"ID":"d40394f2-6a98-49df-813a-17db4ced9cd2","Type":"ContainerDied","Data":"9310cf8e52799cd9ffd94d046c01ec36de9b4fb21b64eff04510a5c86bb2a2cd"} Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.964654 4906 scope.go:117] "RemoveContainer" containerID="9310cf8e52799cd9ffd94d046c01ec36de9b4fb21b64eff04510a5c86bb2a2cd" Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.968516 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-54gvc"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.984913 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-54gvc"] Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990179 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-sb\") pod \"d40394f2-6a98-49df-813a-17db4ced9cd2\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990232 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-nb\") pod \"d40394f2-6a98-49df-813a-17db4ced9cd2\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990303 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-scripts\") pod \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990326 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxxmj\" (UniqueName: \"kubernetes.io/projected/d40394f2-6a98-49df-813a-17db4ced9cd2-kube-api-access-wxxmj\") pod \"d40394f2-6a98-49df-813a-17db4ced9cd2\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990346 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovs-rundir\") pod \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990374 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-metrics-certs-tls-certs\") pod \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990406 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run\") pod \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990465 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67qwj\" (UniqueName: \"kubernetes.io/projected/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-kube-api-access-67qwj\") pod \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990493 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-log-ovn\") pod \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990523 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-ovn-controller-tls-certs\") pod \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990565 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-config\") pod \"d40394f2-6a98-49df-813a-17db4ced9cd2\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990583 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovn-rundir\") pod \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990640 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88cqt\" (UniqueName: \"kubernetes.io/projected/3acf9ff7-0012-4e1a-b940-037bebcc4c88-kube-api-access-88cqt\") pod \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990664 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3acf9ff7-0012-4e1a-b940-037bebcc4c88-config\") pod \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990700 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-swift-storage-0\") pod \"d40394f2-6a98-49df-813a-17db4ced9cd2\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990733 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-svc\") pod \"d40394f2-6a98-49df-813a-17db4ced9cd2\" (UID: \"d40394f2-6a98-49df-813a-17db4ced9cd2\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990773 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run-ovn\") pod \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990788 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-combined-ca-bundle\") pod \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\" (UID: \"3acf9ff7-0012-4e1a-b940-037bebcc4c88\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.990818 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-combined-ca-bundle\") pod \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\" (UID: \"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd\") " Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.995741 4906 generic.go:334] "Generic (PLEG): container finished" podID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerID="9488749cc23de30cd7d4a1998d3ecd858dad70cda1dd749aa047283ebdd924bf" exitCode=143 Nov 23 07:13:43 crc kubenswrapper[4906]: I1123 07:13:43.995911 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5584fdbb5-9b5k8" event={"ID":"2d3c7e43-fe35-428c-99ae-35245b5d62b6","Type":"ContainerDied","Data":"9488749cc23de30cd7d4a1998d3ecd858dad70cda1dd749aa047283ebdd924bf"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.000569 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run" (OuterVolumeSpecName: "var-run") pod "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" (UID: "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.001323 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3acf9ff7-0012-4e1a-b940-037bebcc4c88-config" (OuterVolumeSpecName: "config") pod "3acf9ff7-0012-4e1a-b940-037bebcc4c88" (UID: "3acf9ff7-0012-4e1a-b940-037bebcc4c88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.001396 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" (UID: "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.001428 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" (UID: "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.001461 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "3acf9ff7-0012-4e1a-b940-037bebcc4c88" (UID: "3acf9ff7-0012-4e1a-b940-037bebcc4c88"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.009760 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-scripts" (OuterVolumeSpecName: "scripts") pod "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" (UID: "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.010099 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "3acf9ff7-0012-4e1a-b940-037bebcc4c88" (UID: "3acf9ff7-0012-4e1a-b940-037bebcc4c88"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.011140 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l9d4x" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.011607 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l9d4x" event={"ID":"1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd","Type":"ContainerDied","Data":"031a9b7cf5969785a11c06115f5c7975fef7a3f132b1990849b16ed8da9f3345"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.013052 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.013637 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="1ab3c4dd-5b70-4127-aa8d-570ffcde7477" containerName="nova-cell0-conductor-conductor" containerID="cri-o://a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33" gracePeriod=30 Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.029609 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40394f2-6a98-49df-813a-17db4ced9cd2-kube-api-access-wxxmj" (OuterVolumeSpecName: "kube-api-access-wxxmj") pod "d40394f2-6a98-49df-813a-17db4ced9cd2" (UID: "d40394f2-6a98-49df-813a-17db4ced9cd2"). InnerVolumeSpecName "kube-api-access-wxxmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.030420 4906 generic.go:334] "Generic (PLEG): container finished" podID="4aedece6-346c-4003-8d00-95509a35f23b" containerID="78e2be376a6ac18267e9fa24359c2479618c77eeaf00a6b4b20186ab50f161d6" exitCode=143 Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.030573 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bdb7fc4cb-h962s" event={"ID":"4aedece6-346c-4003-8d00-95509a35f23b","Type":"ContainerDied","Data":"78e2be376a6ac18267e9fa24359c2479618c77eeaf00a6b4b20186ab50f161d6"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.034782 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3acf9ff7-0012-4e1a-b940-037bebcc4c88-kube-api-access-88cqt" (OuterVolumeSpecName: "kube-api-access-88cqt") pod "3acf9ff7-0012-4e1a-b940-037bebcc4c88" (UID: "3acf9ff7-0012-4e1a-b940-037bebcc4c88"). InnerVolumeSpecName "kube-api-access-88cqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.035734 4906 generic.go:334] "Generic (PLEG): container finished" podID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerID="9c4e48b39c14772f337cfc58ea53a79082fd8c022fe4b10170edefdfb0a64859" exitCode=143 Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.035772 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-kube-api-access-67qwj" (OuterVolumeSpecName: "kube-api-access-67qwj") pod "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" (UID: "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd"). InnerVolumeSpecName "kube-api-access-67qwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.035836 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d","Type":"ContainerDied","Data":"9c4e48b39c14772f337cfc58ea53a79082fd8c022fe4b10170edefdfb0a64859"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.038151 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-njmjc_3acf9ff7-0012-4e1a-b940-037bebcc4c88/openstack-network-exporter/0.log" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.038324 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-njmjc" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.039114 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-njmjc" event={"ID":"3acf9ff7-0012-4e1a-b940-037bebcc4c88","Type":"ContainerDied","Data":"b1f9d156bc38ee618a38c1d67bd224d7161e43b996c174f5ccf1865ea86ef638"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.044619 4906 generic.go:334] "Generic (PLEG): container finished" podID="1642db9a-a07d-475a-8112-518d26323d11" containerID="85604e28eb73da42df364a73f2a404a0c2cac5adb3cdbd3da52bb12fba8bccc0" exitCode=143 Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.044759 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1642db9a-a07d-475a-8112-518d26323d11","Type":"ContainerDied","Data":"85604e28eb73da42df364a73f2a404a0c2cac5adb3cdbd3da52bb12fba8bccc0"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.047088 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="0a95e1df-3578-4430-b0ac-befb54723517" containerName="galera" containerID="cri-o://8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85" gracePeriod=30 Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.048311 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d96cfa1a-48d0-4efd-8594-4386a5d7ef35/ovsdbserver-sb/0.log" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.048364 4906 generic.go:334] "Generic (PLEG): container finished" podID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerID="a901c2249d99c2e7c335898965028b8ed5f7a9b34574ccbb6be133b51490d695" exitCode=2 Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.048382 4906 generic.go:334] "Generic (PLEG): container finished" podID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerID="c10b96182b460a717b72550d27ff7aae433f2b3f9e4bf71a704dc8b41e9653fb" exitCode=143 Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.048411 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d96cfa1a-48d0-4efd-8594-4386a5d7ef35","Type":"ContainerDied","Data":"a901c2249d99c2e7c335898965028b8ed5f7a9b34574ccbb6be133b51490d695"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.048447 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d96cfa1a-48d0-4efd-8594-4386a5d7ef35","Type":"ContainerDied","Data":"c10b96182b460a717b72550d27ff7aae433f2b3f9e4bf71a704dc8b41e9653fb"} Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.051565 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2jcp"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094618 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67qwj\" (UniqueName: \"kubernetes.io/projected/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-kube-api-access-67qwj\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094669 4906 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094699 4906 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094709 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88cqt\" (UniqueName: \"kubernetes.io/projected/3acf9ff7-0012-4e1a-b940-037bebcc4c88-kube-api-access-88cqt\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094721 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3acf9ff7-0012-4e1a-b940-037bebcc4c88-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094730 4906 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094739 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094752 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxxmj\" (UniqueName: \"kubernetes.io/projected/d40394f2-6a98-49df-813a-17db4ced9cd2-kube-api-access-wxxmj\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094761 4906 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3acf9ff7-0012-4e1a-b940-037bebcc4c88-ovs-rundir\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.094772 4906 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-var-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.105429 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s2jcp"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.112278 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.148796 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3acf9ff7-0012-4e1a-b940-037bebcc4c88" (UID: "3acf9ff7-0012-4e1a-b940-037bebcc4c88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.197861 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfzfv\" (UniqueName: \"kubernetes.io/projected/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-kube-api-access-rfzfv\") pod \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.198485 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config\") pod \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.198770 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-combined-ca-bundle\") pod \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.198987 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config-secret\") pod \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\" (UID: \"5d56c7ca-9ec3-4cad-bd10-f68eb4339e05\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.199898 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.261275 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-kube-api-access-rfzfv" (OuterVolumeSpecName: "kube-api-access-rfzfv") pod "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" (UID: "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05"). InnerVolumeSpecName "kube-api-access-rfzfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.298932 4906 scope.go:117] "RemoveContainer" containerID="a22234cdf3c895d8bd75acf98c4f78605b788dfa3db4b1b07efc828a4980c561" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.302385 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfzfv\" (UniqueName: \"kubernetes.io/projected/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-kube-api-access-rfzfv\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.311940 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" (UID: "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.346201 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-config" (OuterVolumeSpecName: "config") pod "d40394f2-6a98-49df-813a-17db4ced9cd2" (UID: "d40394f2-6a98-49df-813a-17db4ced9cd2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.408160 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.411113 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.426090 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" (UID: "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.420720 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" (UID: "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.492066 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d40394f2-6a98-49df-813a-17db4ced9cd2" (UID: "d40394f2-6a98-49df-813a-17db4ced9cd2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.514018 4906 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.514439 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.514451 4906 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.523834 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d40394f2-6a98-49df-813a-17db4ced9cd2" (UID: "d40394f2-6a98-49df-813a-17db4ced9cd2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.579106 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d40394f2-6a98-49df-813a-17db4ced9cd2" (UID: "d40394f2-6a98-49df-813a-17db4ced9cd2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.593863 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "3acf9ff7-0012-4e1a-b940-037bebcc4c88" (UID: "3acf9ff7-0012-4e1a-b940-037bebcc4c88"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.595146 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" (UID: "5d56c7ca-9ec3-4cad-bd10-f68eb4339e05"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.601053 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d40394f2-6a98-49df-813a-17db4ced9cd2" (UID: "d40394f2-6a98-49df-813a-17db4ced9cd2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.617158 4906 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.617192 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.617202 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d40394f2-6a98-49df-813a-17db4ced9cd2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.617214 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3acf9ff7-0012-4e1a-b940-037bebcc4c88-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.617228 4906 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.713732 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder46ad-account-delete-rpn4p"] Nov 23 07:13:44 crc kubenswrapper[4906]: E1123 07:13:44.732248 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:44 crc kubenswrapper[4906]: E1123 07:13:44.732354 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data podName:4d677a57-1acf-4627-aa66-f4bee96a2b51 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:48.732318992 +0000 UTC m=+1444.245710295 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data") pod "rabbitmq-cell1-server-0" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51") : configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.742966 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d96cfa1a-48d0-4efd-8594-4386a5d7ef35/ovsdbserver-sb/0.log" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.743098 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.775109 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" (UID: "1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.782863 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementb64f-account-delete-576n6"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.804339 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicana450-account-delete-l2bs5"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.830118 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glanceaaf0-account-delete-j59k7"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.840396 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-njmjc"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.845989 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-metrics-certs-tls-certs\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.846056 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-config\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.846350 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-combined-ca-bundle\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.846403 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdb-rundir\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.846506 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-scripts\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.846651 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.846720 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pk52\" (UniqueName: \"kubernetes.io/projected/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-kube-api-access-5pk52\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.846740 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdbserver-sb-tls-certs\") pod \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\" (UID: \"d96cfa1a-48d0-4efd-8594-4386a5d7ef35\") " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.847368 4906 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.848191 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-njmjc"] Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.849031 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-scripts" (OuterVolumeSpecName: "scripts") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.849110 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.849625 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-config" (OuterVolumeSpecName: "config") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.854230 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.855308 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-kube-api-access-5pk52" (OuterVolumeSpecName: "kube-api-access-5pk52") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "kube-api-access-5pk52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.903628 4906 scope.go:117] "RemoveContainer" containerID="5bfedeaefcd18a2c6080a3fb6a22a56ae3610931424a75b1dde46d3765c0a44e" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.926596 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.935918 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.950250 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.950379 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.950394 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.950407 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pk52\" (UniqueName: \"kubernetes.io/projected/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-kube-api-access-5pk52\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.950417 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.950425 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.950434 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:44 crc kubenswrapper[4906]: I1123 07:13:44.983650 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:44.999275 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "d96cfa1a-48d0-4efd-8594-4386a5d7ef35" (UID: "d96cfa1a-48d0-4efd-8594-4386a5d7ef35"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.016728 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l9d4x"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.021382 4906 scope.go:117] "RemoveContainer" containerID="28d9044f3cc3149e597f205efe1d60660461a96061a5b3c9310711ba4008ce47" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.025244 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-l9d4x"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.047413 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.054021 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.054121 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d96cfa1a-48d0-4efd-8594-4386a5d7ef35-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.068083 4906 generic.go:334] "Generic (PLEG): container finished" podID="b4121204-6477-44c6-b669-fc7497fa908d" containerID="dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113" exitCode=0 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.068122 4906 generic.go:334] "Generic (PLEG): container finished" podID="b4121204-6477-44c6-b669-fc7497fa908d" containerID="c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7" exitCode=0 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.068175 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77c4d866c-r9x72" event={"ID":"b4121204-6477-44c6-b669-fc7497fa908d","Type":"ContainerDied","Data":"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.068241 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77c4d866c-r9x72" event={"ID":"b4121204-6477-44c6-b669-fc7497fa908d","Type":"ContainerDied","Data":"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.068257 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-77c4d866c-r9x72" event={"ID":"b4121204-6477-44c6-b669-fc7497fa908d","Type":"ContainerDied","Data":"e6db7921dc7dbf6ed9d3d94c5afccf6bd7f05007a7f4deb53dcdac820859140f"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.068345 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-77c4d866c-r9x72" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.069491 4906 scope.go:117] "RemoveContainer" containerID="dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.083553 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" event={"ID":"d40394f2-6a98-49df-813a-17db4ced9cd2","Type":"ContainerDied","Data":"efd52e7666be2f2f4da0cc1d0490f9a3f79d80bbbd90cea212bdde53a11e6688"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.083736 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.093641 4906 generic.go:334] "Generic (PLEG): container finished" podID="6cee328a-b746-486d-850b-148ca1e66eea" containerID="18d1eb931158c01bd8771394295d1f098c58d8673d4374a70905fb56a217df1f" exitCode=0 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.093798 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cee328a-b746-486d-850b-148ca1e66eea","Type":"ContainerDied","Data":"18d1eb931158c01bd8771394295d1f098c58d8673d4374a70905fb56a217df1f"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.098301 4906 generic.go:334] "Generic (PLEG): container finished" podID="0a95e1df-3578-4430-b0ac-befb54723517" containerID="8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85" exitCode=0 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.098399 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0a95e1df-3578-4430-b0ac-befb54723517","Type":"ContainerDied","Data":"8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.100741 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d96cfa1a-48d0-4efd-8594-4386a5d7ef35/ovsdbserver-sb/0.log" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.100828 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d96cfa1a-48d0-4efd-8594-4386a5d7ef35","Type":"ContainerDied","Data":"4dab160dbcb6dc297117e7a0680bd44a1741500bb00c62e450d050c5c4b6ccb3"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.100930 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.102899 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceaaf0-account-delete-j59k7" event={"ID":"2e0ab9b5-09c4-49b9-a112-76c18f06524d","Type":"ContainerStarted","Data":"d0f6fec4ee2f6b122dca6c37a54aed3f3e5cc1be47767f587d23965550258531"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.105777 4906 generic.go:334] "Generic (PLEG): container finished" podID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" exitCode=0 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.105875 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ftw6h" event={"ID":"e34ddce9-d32e-4c31-9716-86ad39a590e9","Type":"ContainerDied","Data":"cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.116672 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder46ad-account-delete-rpn4p" event={"ID":"4df634a0-630f-46a5-b304-86211157136a","Type":"ContainerStarted","Data":"55bcf6a4b1677ea5ec89ebefc19f137bea7a0453cd26cf7bf4cc78e4e7d235ec"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.127386 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicana450-account-delete-l2bs5" event={"ID":"64c1c989-3443-4b80-9786-56b68660655c","Type":"ContainerStarted","Data":"65f008398b0a61b9837f9c35e83f9327847033e6ccd77632fde49a148a93fb1c"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.139844 4906 generic.go:334] "Generic (PLEG): container finished" podID="e84af371-9517-42aa-943b-a4847f4f8577" containerID="5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390" exitCode=0 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.139943 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f84fc787f-zskfn" event={"ID":"e84af371-9517-42aa-943b-a4847f4f8577","Type":"ContainerDied","Data":"5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155125 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-log-httpd\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155207 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-etc-swift\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155238 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-run-httpd\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155277 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-combined-ca-bundle\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155369 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfrqx\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-kube-api-access-jfrqx\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155556 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-internal-tls-certs\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155747 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-config-data\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155804 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-public-tls-certs\") pod \"b4121204-6477-44c6-b669-fc7497fa908d\" (UID: \"b4121204-6477-44c6-b669-fc7497fa908d\") " Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.156376 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.156876 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.158485 4906 generic.go:334] "Generic (PLEG): container finished" podID="00d152f2-1543-4b14-b214-73747a3f9a26" containerID="cfe0d8e3ca16e113f2016b61cb50f6b5b415f288a07bd395effcd5f6002c7ed6" exitCode=143 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.158615 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00d152f2-1543-4b14-b214-73747a3f9a26","Type":"ContainerDied","Data":"cfe0d8e3ca16e113f2016b61cb50f6b5b415f288a07bd395effcd5f6002c7ed6"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.155371 4906 scope.go:117] "RemoveContainer" containerID="c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.166012 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.166049 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4121204-6477-44c6-b669-fc7497fa908d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.190887 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.196184 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-kube-api-access-jfrqx" (OuterVolumeSpecName: "kube-api-access-jfrqx") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "kube-api-access-jfrqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.199576 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.224008 4906 generic.go:334] "Generic (PLEG): container finished" podID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerID="d7113d47184e50909b0d719000d22c051debd5b0dca75d3f957a7612dc3670a1" exitCode=143 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.224405 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d7b2888-eaed-4879-9f60-e4264aed2287","Type":"ContainerDied","Data":"d7113d47184e50909b0d719000d22c051debd5b0dca75d3f957a7612dc3670a1"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.226787 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-944h6"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.255594 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-944h6"] Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.258786 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0 is running failed: container process not found" containerID="2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.269037 4906 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.269071 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfrqx\" (UniqueName: \"kubernetes.io/projected/b4121204-6477-44c6-b669-fc7497fa908d-kube-api-access-jfrqx\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.269358 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0 is running failed: container process not found" containerID="2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.269473 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.269999 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0 is running failed: container process not found" containerID="2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.270072 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="ovsdbserver-nb" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.272157 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementb64f-account-delete-576n6" event={"ID":"6093f84a-c555-4fc4-a207-f54e5ddf2a85","Type":"ContainerStarted","Data":"e161829b4c3412d3950ffec321774284108121997a6f8f59cc52bdcd36b4f608"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.277351 4906 generic.go:334] "Generic (PLEG): container finished" podID="a7788825-483c-46f2-a975-239254435019" containerID="9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c" exitCode=0 Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.277426 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7788825-483c-46f2-a975-239254435019","Type":"ContainerDied","Data":"9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c"} Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.277945 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.309306 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron9f60-account-delete-vq457"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.329448 4906 scope.go:117] "RemoveContainer" containerID="dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113" Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.330250 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113\": container with ID starting with dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113 not found: ID does not exist" containerID="dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.330432 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113"} err="failed to get container status \"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113\": rpc error: code = NotFound desc = could not find container \"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113\": container with ID starting with dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113 not found: ID does not exist" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.330512 4906 scope.go:117] "RemoveContainer" containerID="c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7" Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.333102 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7\": container with ID starting with c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7 not found: ID does not exist" containerID="c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.333165 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7"} err="failed to get container status \"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7\": rpc error: code = NotFound desc = could not find container \"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7\": container with ID starting with c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7 not found: ID does not exist" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.333205 4906 scope.go:117] "RemoveContainer" containerID="dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.334782 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113"} err="failed to get container status \"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113\": rpc error: code = NotFound desc = could not find container \"dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113\": container with ID starting with dadde181875d4518ca44152a01f42b96007c67e1258b8c5c2adf1e44667ae113 not found: ID does not exist" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.334873 4906 scope.go:117] "RemoveContainer" containerID="c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.338020 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7"} err="failed to get container status \"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7\": rpc error: code = NotFound desc = could not find container \"c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7\": container with ID starting with c360be6417dbb8c084b61aa88001b17ea3819909dd9bc29ee2c34bfbfed3d9b7 not found: ID does not exist" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.338107 4906 scope.go:117] "RemoveContainer" containerID="a901c2249d99c2e7c335898965028b8ed5f7a9b34574ccbb6be133b51490d695" Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.370390 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.370448 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data podName:9a0222ea-7767-4d08-a6ec-6659a33f9df2 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:49.370435023 +0000 UTC m=+1444.883826326 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data") pod "rabbitmq-server-0" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2") : configmap "rabbitmq-config-data" not found Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.373396 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03f12dad-1cb2-4e43-b9cc-7658ed44bc1b" path="/var/lib/kubelet/pods/03f12dad-1cb2-4e43-b9cc-7658ed44bc1b/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.374019 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" path="/var/lib/kubelet/pods/1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.375944 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3acf9ff7-0012-4e1a-b940-037bebcc4c88" path="/var/lib/kubelet/pods/3acf9ff7-0012-4e1a-b940-037bebcc4c88/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.377162 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d56c7ca-9ec3-4cad-bd10-f68eb4339e05" path="/var/lib/kubelet/pods/5d56c7ca-9ec3-4cad-bd10-f68eb4339e05/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.380441 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a2c54f8-cc51-4e33-9b70-62107c26f267" path="/var/lib/kubelet/pods/7a2c54f8-cc51-4e33-9b70-62107c26f267/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.386559 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7be07491-e9f1-4b75-9261-c3f49dacb542" path="/var/lib/kubelet/pods/7be07491-e9f1-4b75-9261-c3f49dacb542/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: W1123 07:13:45.387669 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00991da6_745e_4757_9d89_a20c5b97e38f.slice/crio-ed471134383b6a342d71e4523be2b2493de4106ca7705b9219d7f8842e8521cc WatchSource:0}: Error finding container ed471134383b6a342d71e4523be2b2493de4106ca7705b9219d7f8842e8521cc: Status 404 returned error can't find the container with id ed471134383b6a342d71e4523be2b2493de4106ca7705b9219d7f8842e8521cc Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.396968 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" path="/var/lib/kubelet/pods/d40394f2-6a98-49df-813a-17db4ced9cd2/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.397816 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" path="/var/lib/kubelet/pods/d96cfa1a-48d0-4efd-8594-4386a5d7ef35/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.416350 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df30375c-b83c-490c-be3d-2453779a3c06" path="/var/lib/kubelet/pods/df30375c-b83c-490c-be3d-2453779a3c06/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.421460 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc92acb2-4da0-4ff3-8489-e63fb755861b" path="/var/lib/kubelet/pods/fc92acb2-4da0-4ff3-8489-e63fb755861b/volumes" Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.589455 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85 is running failed: container process not found" containerID="8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.592020 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85 is running failed: container process not found" containerID="8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.592354 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85 is running failed: container process not found" containerID="8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 23 07:13:45 crc kubenswrapper[4906]: E1123 07:13:45.592395 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85 is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="0a95e1df-3578-4430-b0ac-befb54723517" containerName="galera" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.661909 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-config-data" (OuterVolumeSpecName: "config-data") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.665099 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.678266 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.679362 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b4121204-6477-44c6-b669-fc7497fa908d" (UID: "b4121204-6477-44c6-b669-fc7497fa908d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.682581 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.682609 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.682621 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.682631 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4121204-6477-44c6-b669-fc7497fa908d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.847041 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi944e-account-delete-gw8gk"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.847478 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell082ff-account-delete-s8jc5"] Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.885814 4906 scope.go:117] "RemoveContainer" containerID="c10b96182b460a717b72550d27ff7aae433f2b3f9e4bf71a704dc8b41e9653fb" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.894408 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_51dd433b-f1a2-422a-bf5b-643e150c3bca/ovsdbserver-nb/0.log" Nov 23 07:13:45 crc kubenswrapper[4906]: I1123 07:13:45.894608 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.000430 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-combined-ca-bundle\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.000838 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdbserver-nb-tls-certs\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.001721 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-metrics-certs-tls-certs\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.003819 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.004567 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-config\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.005004 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5zkc\" (UniqueName: \"kubernetes.io/projected/51dd433b-f1a2-422a-bf5b-643e150c3bca-kube-api-access-r5zkc\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.006602 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-scripts\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.006762 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdb-rundir\") pod \"51dd433b-f1a2-422a-bf5b-643e150c3bca\" (UID: \"51dd433b-f1a2-422a-bf5b-643e150c3bca\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.008087 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-config" (OuterVolumeSpecName: "config") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.008158 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-scripts" (OuterVolumeSpecName: "scripts") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.008339 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.010769 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51dd433b-f1a2-422a-bf5b-643e150c3bca-kube-api-access-r5zkc" (OuterVolumeSpecName: "kube-api-access-r5zkc") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "kube-api-access-r5zkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.021970 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.072892 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.093454 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-77c4d866c-r9x72"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.111994 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-77c4d866c-r9x72"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.115247 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.115410 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-combined-ca-bundle\") pod \"6cee328a-b746-486d-850b-148ca1e66eea\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.115534 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mp7r\" (UniqueName: \"kubernetes.io/projected/6cee328a-b746-486d-850b-148ca1e66eea-kube-api-access-4mp7r\") pod \"6cee328a-b746-486d-850b-148ca1e66eea\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.115821 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-config-data\") pod \"6cee328a-b746-486d-850b-148ca1e66eea\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.115912 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.116009 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-nova-novncproxy-tls-certs\") pod \"6cee328a-b746-486d-850b-148ca1e66eea\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.116116 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-vencrypt-tls-certs\") pod \"6cee328a-b746-486d-850b-148ca1e66eea\" (UID: \"6cee328a-b746-486d-850b-148ca1e66eea\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.117012 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.118813 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.118945 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5zkc\" (UniqueName: \"kubernetes.io/projected/51dd433b-f1a2-422a-bf5b-643e150c3bca-kube-api-access-r5zkc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.119005 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51dd433b-f1a2-422a-bf5b-643e150c3bca-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.119055 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.128362 4906 scope.go:117] "RemoveContainer" containerID="6bc391c38da2e27698e0bb15d809a3a29934990091bf87fa84600dc4d231301b" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.135980 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cee328a-b746-486d-850b-148ca1e66eea-kube-api-access-4mp7r" (OuterVolumeSpecName: "kube-api-access-4mp7r") pod "6cee328a-b746-486d-850b-148ca1e66eea" (UID: "6cee328a-b746-486d-850b-148ca1e66eea"). InnerVolumeSpecName "kube-api-access-4mp7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.220708 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-galera-tls-certs\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.220793 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxt8r\" (UniqueName: \"kubernetes.io/projected/a7788825-483c-46f2-a975-239254435019-kube-api-access-jxt8r\") pod \"a7788825-483c-46f2-a975-239254435019\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.220845 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.220902 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-combined-ca-bundle\") pod \"a7788825-483c-46f2-a975-239254435019\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.220936 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7788825-483c-46f2-a975-239254435019-etc-machine-id\") pod \"a7788825-483c-46f2-a975-239254435019\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.220962 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data\") pod \"a7788825-483c-46f2-a975-239254435019\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.220994 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0a95e1df-3578-4430-b0ac-befb54723517-config-data-generated\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221054 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-combined-ca-bundle\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221122 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data-custom\") pod \"a7788825-483c-46f2-a975-239254435019\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221176 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgbzq\" (UniqueName: \"kubernetes.io/projected/0a95e1df-3578-4430-b0ac-befb54723517-kube-api-access-zgbzq\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221208 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-operator-scripts\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221210 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7788825-483c-46f2-a975-239254435019-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a7788825-483c-46f2-a975-239254435019" (UID: "a7788825-483c-46f2-a975-239254435019"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221284 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-config-data-default\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221308 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-scripts\") pod \"a7788825-483c-46f2-a975-239254435019\" (UID: \"a7788825-483c-46f2-a975-239254435019\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221332 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-kolla-config\") pod \"0a95e1df-3578-4430-b0ac-befb54723517\" (UID: \"0a95e1df-3578-4430-b0ac-befb54723517\") " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221879 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mp7r\" (UniqueName: \"kubernetes.io/projected/6cee328a-b746-486d-850b-148ca1e66eea-kube-api-access-4mp7r\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.221909 4906 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7788825-483c-46f2-a975-239254435019-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.222483 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.223196 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.223632 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.224239 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a95e1df-3578-4430-b0ac-befb54723517-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.252483 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a7788825-483c-46f2-a975-239254435019" (UID: "a7788825-483c-46f2-a975-239254435019"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.255892 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-scripts" (OuterVolumeSpecName: "scripts") pod "a7788825-483c-46f2-a975-239254435019" (UID: "a7788825-483c-46f2-a975-239254435019"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.273650 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7788825-483c-46f2-a975-239254435019-kube-api-access-jxt8r" (OuterVolumeSpecName: "kube-api-access-jxt8r") pod "a7788825-483c-46f2-a975-239254435019" (UID: "a7788825-483c-46f2-a975-239254435019"). InnerVolumeSpecName "kube-api-access-jxt8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.276487 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a95e1df-3578-4430-b0ac-befb54723517-kube-api-access-zgbzq" (OuterVolumeSpecName: "kube-api-access-zgbzq") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "kube-api-access-zgbzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.307182 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceaaf0-account-delete-j59k7" event={"ID":"2e0ab9b5-09c4-49b9-a112-76c18f06524d","Type":"ContainerStarted","Data":"abee71bb44307a2c4573a7acf7731e1011f5ab2e46aced3873ac6dfb10570fe3"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.316467 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.316766 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cee328a-b746-486d-850b-148ca1e66eea","Type":"ContainerDied","Data":"0defa47ee7111204fd4f55204c7ca6efa62d6a16a0d4e3e0329a19e00784a029"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.316841 4906 scope.go:117] "RemoveContainer" containerID="18d1eb931158c01bd8771394295d1f098c58d8673d4374a70905fb56a217df1f" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.325969 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0a95e1df-3578-4430-b0ac-befb54723517-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.325998 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.326008 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgbzq\" (UniqueName: \"kubernetes.io/projected/0a95e1df-3578-4430-b0ac-befb54723517-kube-api-access-zgbzq\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.326018 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.326028 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.326038 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.326048 4906 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a95e1df-3578-4430-b0ac-befb54723517-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.326061 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxt8r\" (UniqueName: \"kubernetes.io/projected/a7788825-483c-46f2-a975-239254435019-kube-api-access-jxt8r\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.336996 4906 generic.go:334] "Generic (PLEG): container finished" podID="da74c07b-2832-4d8e-9508-415fdc3cac71" containerID="4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f" exitCode=0 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.337126 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da74c07b-2832-4d8e-9508-415fdc3cac71","Type":"ContainerDied","Data":"4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.342581 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glanceaaf0-account-delete-j59k7" podStartSLOduration=5.342564485 podStartE2EDuration="5.342564485s" podCreationTimestamp="2025-11-23 07:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:13:46.336778983 +0000 UTC m=+1441.850170286" watchObservedRunningTime="2025-11-23 07:13:46.342564485 +0000 UTC m=+1441.855955788" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.353071 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.363485 4906 generic.go:334] "Generic (PLEG): container finished" podID="a7788825-483c-46f2-a975-239254435019" containerID="209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d" exitCode=0 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.363867 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7788825-483c-46f2-a975-239254435019","Type":"ContainerDied","Data":"209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.363944 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a7788825-483c-46f2-a975-239254435019","Type":"ContainerDied","Data":"55ac9528202f43bf2d8a88071468bd7753abea70b0cbb303b4d4b35b5acac902"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.363989 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.369534 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi944e-account-delete-gw8gk" event={"ID":"0ab42792-22e3-48ee-8028-3d181679190c","Type":"ContainerStarted","Data":"d56f185ae462d261cefb6e5dbaf0ecb893e6dedd267291f9b4fa4cdb06b06e4b"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.372310 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.373144 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.374622 4906 generic.go:334] "Generic (PLEG): container finished" podID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerID="e5fc6b77a408f81d5b91fd45ba27278acb2c5d247b602dbc2475870c5ba45b81" exitCode=0 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.374672 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ffb7cbc86-lz7bc" event={"ID":"73149d5c-c16f-4b83-b855-135fc1aab9fe","Type":"ContainerDied","Data":"e5fc6b77a408f81d5b91fd45ba27278acb2c5d247b602dbc2475870c5ba45b81"} Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.389107 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.392779 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.395790 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.395828 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="212ee22b-c3e7-4d9c-b952-5837b5982c63" containerName="nova-cell1-conductor-conductor" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.396177 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron9f60-account-delete-vq457" event={"ID":"00991da6-745e-4757-9d89-a20c5b97e38f","Type":"ContainerStarted","Data":"ed471134383b6a342d71e4523be2b2493de4106ca7705b9219d7f8842e8521cc"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.403865 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell082ff-account-delete-s8jc5" event={"ID":"01d907ea-d712-40af-b18a-5d55287fbaeb","Type":"ContainerStarted","Data":"95317f08311eec38eda1ec3bca3eb11cc85b1f3aeb23b3c83a54d447ed5285e0"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.412318 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.412308 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0a95e1df-3578-4430-b0ac-befb54723517","Type":"ContainerDied","Data":"b27a842c89676241116fe33499f637397fd007b238dd4c2a6f1c1470abbc6662"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.419672 4906 generic.go:334] "Generic (PLEG): container finished" podID="4df634a0-630f-46a5-b304-86211157136a" containerID="f7be6a395372de3ef443ddccc155ec0954c9663ac6663a6358c697025786322a" exitCode=0 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.419785 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder46ad-account-delete-rpn4p" event={"ID":"4df634a0-630f-46a5-b304-86211157136a","Type":"ContainerDied","Data":"f7be6a395372de3ef443ddccc155ec0954c9663ac6663a6358c697025786322a"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.423903 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron9f60-account-delete-vq457" podStartSLOduration=5.423887215 podStartE2EDuration="5.423887215s" podCreationTimestamp="2025-11-23 07:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:13:46.414250193 +0000 UTC m=+1441.927641496" watchObservedRunningTime="2025-11-23 07:13:46.423887215 +0000 UTC m=+1441.937278518" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.428523 4906 generic.go:334] "Generic (PLEG): container finished" podID="6093f84a-c555-4fc4-a207-f54e5ddf2a85" containerID="8da666a9e40465436d1857a8452785646b2c0c473e10ad242db4541b87e6a037" exitCode=0 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.428590 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementb64f-account-delete-576n6" event={"ID":"6093f84a-c555-4fc4-a207-f54e5ddf2a85","Type":"ContainerDied","Data":"8da666a9e40465436d1857a8452785646b2c0c473e10ad242db4541b87e6a037"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.429251 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.429299 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.429311 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.439562 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_51dd433b-f1a2-422a-bf5b-643e150c3bca/ovsdbserver-nb/0.log" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.439662 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"51dd433b-f1a2-422a-bf5b-643e150c3bca","Type":"ContainerDied","Data":"56e048bf7af7ca4d54382a670d2c0118e5ad996cb59a2692751315f09133d3a8"} Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.439804 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.511525 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.161:8776/healthcheck\": read tcp 10.217.0.2:44218->10.217.0.161:8776: read: connection reset by peer" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.657004 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.657363 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-central-agent" containerID="cri-o://f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df" gracePeriod=30 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.657857 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="proxy-httpd" containerID="cri-o://4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0" gracePeriod=30 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.657901 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="sg-core" containerID="cri-o://009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5" gracePeriod=30 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.657948 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-notification-agent" containerID="cri-o://f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f" gracePeriod=30 Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.677025 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f is running failed: container process not found" containerID="4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.678842 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.679076 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9535028a-afb6-49d6-9bac-541c5324cc55" containerName="kube-state-metrics" containerID="cri-o://0ef240d12f0659c3ea6a9b32c942f26f9370dad6938a45b1af45fb31d908312f" gracePeriod=30 Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.682402 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f is running failed: container process not found" containerID="4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.686806 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f is running failed: container process not found" containerID="4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.686910 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="da74c07b-2832-4d8e-9508-415fdc3cac71" containerName="nova-scheduler-scheduler" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.746772 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.776150 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.779854 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.781641 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.781709 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.798608 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.808981 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.813999 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.814086 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.842502 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.843973 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="27d994c3-4d19-4084-8c09-d0f2a8d7ede7" containerName="memcached" containerID="cri-o://c62d53e7d032049f41cdae0eac7ed993f6453996187bc0ab5a5c3458db844d22" gracePeriod=30 Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.844437 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.872342 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-78bh5"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.883575 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.886094 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.889040 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 23 07:13:46 crc kubenswrapper[4906]: E1123 07:13:46.889109 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="1ab3c4dd-5b70-4127-aa8d-570ffcde7477" containerName="nova-cell0-conductor-conductor" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.913869 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-78bh5"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.935720 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-c7r8h"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.978858 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-c7r8h"] Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.979062 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.173:9292/healthcheck\": read tcp 10.217.0.2:52166->10.217.0.173:9292: read: connection reset by peer" Nov 23 07:13:46 crc kubenswrapper[4906]: I1123 07:13:46.979384 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.173:9292/healthcheck\": read tcp 10.217.0.2:52178->10.217.0.173:9292: read: connection reset by peer" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.004345 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-87fd667c9-tzc42"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.004644 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-87fd667c9-tzc42" podUID="2e303c56-48b8-4176-99f6-fd0aeb26cf94" containerName="keystone-api" containerID="cri-o://4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb" gracePeriod=30 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.016375 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.046870 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-f9bjf"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.059454 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-f9bjf"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.065149 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-92c3-account-create-25d6n"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.087323 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-92c3-account-create-25d6n"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.136562 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cee328a-b746-486d-850b-148ca1e66eea" (UID: "6cee328a-b746-486d-850b-148ca1e66eea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.142963 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bdb7fc4cb-h962s" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:36724->10.217.0.156:9311: read: connection reset by peer" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.143044 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bdb7fc4cb-h962s" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:36722->10.217.0.156:9311: read: connection reset by peer" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.144514 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-config-data" (OuterVolumeSpecName: "config-data") pod "6cee328a-b746-486d-850b-148ca1e66eea" (UID: "6cee328a-b746-486d-850b-148ca1e66eea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.164884 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.174742 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:41196->10.217.0.200:8775: read: connection reset by peer" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.175150 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:41186->10.217.0.200:8775: read: connection reset by peer" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.178582 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.183447 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.183460 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.219729 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.285808 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.328598 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "6cee328a-b746-486d-850b-148ca1e66eea" (UID: "6cee328a-b746-486d-850b-148ca1e66eea"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.341439 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7788825-483c-46f2-a975-239254435019" (UID: "a7788825-483c-46f2-a975-239254435019"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.341465 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "51dd433b-f1a2-422a-bf5b-643e150c3bca" (UID: "51dd433b-f1a2-422a-bf5b-643e150c3bca"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.388346 4906 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.388389 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.388401 4906 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51dd433b-f1a2-422a-bf5b-643e150c3bca-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.389887 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.175:9292/healthcheck\": dial tcp 10.217.0.175:9292: connect: connection refused" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.406143 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.175:9292/healthcheck\": dial tcp 10.217.0.175:9292: connect: connection refused" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.452726 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "6cee328a-b746-486d-850b-148ca1e66eea" (UID: "6cee328a-b746-486d-850b-148ca1e66eea"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.491355 4906 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee328a-b746-486d-850b-148ca1e66eea-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.534985 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data" (OuterVolumeSpecName: "config-data") pod "a7788825-483c-46f2-a975-239254435019" (UID: "a7788825-483c-46f2-a975-239254435019"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.560774 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "0a95e1df-3578-4430-b0ac-befb54723517" (UID: "0a95e1df-3578-4430-b0ac-befb54723517"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.591544 4906 generic.go:334] "Generic (PLEG): container finished" podID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerID="77ce16e41b99ad023af4aa1620f90f3b6cbe24953a476cead47e6c790b48478b" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.600308 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55501b57-eb6b-4753-a9bd-43ff19f6c7a9" path="/var/lib/kubelet/pods/55501b57-eb6b-4753-a9bd-43ff19f6c7a9/volumes" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.601352 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61603290-df20-4ffb-af90-16aef4437224" path="/var/lib/kubelet/pods/61603290-df20-4ffb-af90-16aef4437224/volumes" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.601960 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4121204-6477-44c6-b669-fc7497fa908d" path="/var/lib/kubelet/pods/b4121204-6477-44c6-b669-fc7497fa908d/volumes" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.603070 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7e51ef-70f7-4fb2-a1ef-31a109125528" path="/var/lib/kubelet/pods/bb7e51ef-70f7-4fb2-a1ef-31a109125528/volumes" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.603627 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d86996ce-011e-4d9b-85c0-dae8ba298edd" path="/var/lib/kubelet/pods/d86996ce-011e-4d9b-85c0-dae8ba298edd/volumes" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.604841 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d7b2888-eaed-4879-9f60-e4264aed2287","Type":"ContainerDied","Data":"77ce16e41b99ad023af4aa1620f90f3b6cbe24953a476cead47e6c790b48478b"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.604879 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicana450-account-delete-l2bs5" event={"ID":"64c1c989-3443-4b80-9786-56b68660655c","Type":"ContainerStarted","Data":"bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.604894 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4r2qz"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.604911 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4r2qz"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.609248 4906 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicana450-account-delete-l2bs5" secret="" err="secret \"galera-openstack-dockercfg-gqq6f\" not found" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.611224 4906 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a95e1df-3578-4430-b0ac-befb54723517-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.611268 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7788825-483c-46f2-a975-239254435019-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.633217 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerName="galera" containerID="cri-o://27497e40b508a21677f03a6794d604b1ad821f2140f70be90b3a8b007d2d719d" gracePeriod=30 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.634259 4906 generic.go:334] "Generic (PLEG): container finished" podID="0ab42792-22e3-48ee-8028-3d181679190c" containerID="168ec284586fb5aa74b2d96f311c089e82fd4a4806a8147df56bdd9a2c39ee76" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.634352 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi944e-account-delete-gw8gk" event={"ID":"0ab42792-22e3-48ee-8028-3d181679190c","Type":"ContainerDied","Data":"168ec284586fb5aa74b2d96f311c089e82fd4a4806a8147df56bdd9a2c39ee76"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.654774 4906 generic.go:334] "Generic (PLEG): container finished" podID="9535028a-afb6-49d6-9bac-541c5324cc55" containerID="0ef240d12f0659c3ea6a9b32c942f26f9370dad6938a45b1af45fb31d908312f" exitCode=2 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.654877 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9535028a-afb6-49d6-9bac-541c5324cc55","Type":"ContainerDied","Data":"0ef240d12f0659c3ea6a9b32c942f26f9370dad6938a45b1af45fb31d908312f"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.665047 4906 generic.go:334] "Generic (PLEG): container finished" podID="00d152f2-1543-4b14-b214-73747a3f9a26" containerID="59399848d4fd8c548b0ad9fe61db4277a707a18507241aca88683dc29eebb13a" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.665180 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00d152f2-1543-4b14-b214-73747a3f9a26","Type":"ContainerDied","Data":"59399848d4fd8c548b0ad9fe61db4277a707a18507241aca88683dc29eebb13a"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.672902 4906 generic.go:334] "Generic (PLEG): container finished" podID="2e0ab9b5-09c4-49b9-a112-76c18f06524d" containerID="abee71bb44307a2c4573a7acf7731e1011f5ab2e46aced3873ac6dfb10570fe3" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.672998 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceaaf0-account-delete-j59k7" event={"ID":"2e0ab9b5-09c4-49b9-a112-76c18f06524d","Type":"ContainerDied","Data":"abee71bb44307a2c4573a7acf7731e1011f5ab2e46aced3873ac6dfb10570fe3"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.678710 4906 generic.go:334] "Generic (PLEG): container finished" podID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerID="4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.678770 4906 generic.go:334] "Generic (PLEG): container finished" podID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerID="009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5" exitCode=2 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.680071 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerDied","Data":"4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.680120 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerDied","Data":"009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.684563 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi944e-account-delete-gw8gk"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.685106 4906 generic.go:334] "Generic (PLEG): container finished" podID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerID="2fd16719fcd506b038a0152208792ed142945112e8297ebc99ad99b173d43f11" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.685160 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d","Type":"ContainerDied","Data":"2fd16719fcd506b038a0152208792ed142945112e8297ebc99ad99b173d43f11"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.688892 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell082ff-account-delete-s8jc5" event={"ID":"01d907ea-d712-40af-b18a-5d55287fbaeb","Type":"ContainerStarted","Data":"098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.690310 4906 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell082ff-account-delete-s8jc5" secret="" err="secret \"galera-openstack-dockercfg-gqq6f\" not found" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.693769 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ffb7cbc86-lz7bc" event={"ID":"73149d5c-c16f-4b83-b855-135fc1aab9fe","Type":"ContainerDied","Data":"5bd4727bfb09b7d4d7d82d1343a0b814629a6bdd84d7c99f84522b703b964091"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.693797 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bd4727bfb09b7d4d7d82d1343a0b814629a6bdd84d7c99f84522b703b964091" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.695813 4906 generic.go:334] "Generic (PLEG): container finished" podID="4aedece6-346c-4003-8d00-95509a35f23b" containerID="0dd9b19587494f220b9ce7cb80f137d8caf2f6b1707a1f5bae2b3bc3625d12af" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.695885 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bdb7fc4cb-h962s" event={"ID":"4aedece6-346c-4003-8d00-95509a35f23b","Type":"ContainerDied","Data":"0dd9b19587494f220b9ce7cb80f137d8caf2f6b1707a1f5bae2b3bc3625d12af"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.702518 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da74c07b-2832-4d8e-9508-415fdc3cac71","Type":"ContainerDied","Data":"26285d7556a44fb9a5857cf96431aae6fce54b8ddebef74d066f8afbf62a94b5"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.702581 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26285d7556a44fb9a5857cf96431aae6fce54b8ddebef74d066f8afbf62a94b5" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.705357 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-944e-account-create-zh9ms"] Nov 23 07:13:47 crc kubenswrapper[4906]: E1123 07:13:47.714323 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:47 crc kubenswrapper[4906]: E1123 07:13:47.714388 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts podName:01d907ea-d712-40af-b18a-5d55287fbaeb nodeName:}" failed. No retries permitted until 2025-11-23 07:13:48.214370139 +0000 UTC m=+1443.727761442 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts") pod "novacell082ff-account-delete-s8jc5" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb") : configmap "openstack-scripts" not found Nov 23 07:13:47 crc kubenswrapper[4906]: E1123 07:13:47.714613 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:47 crc kubenswrapper[4906]: E1123 07:13:47.714640 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts podName:64c1c989-3443-4b80-9786-56b68660655c nodeName:}" failed. No retries permitted until 2025-11-23 07:13:48.214632616 +0000 UTC m=+1443.728023919 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts") pod "barbicana450-account-delete-l2bs5" (UID: "64c1c989-3443-4b80-9786-56b68660655c") : configmap "openstack-scripts" not found Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.716480 4906 generic.go:334] "Generic (PLEG): container finished" podID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerID="e82535f708e61c7ee739e897ea48ec09a393ccd402bebffaf681b12cea2dcb54" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.716560 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-944e-account-create-zh9ms"] Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.716591 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ee97571-4ae5-4ed0-a32a-67abd5896ae0","Type":"ContainerDied","Data":"e82535f708e61c7ee739e897ea48ec09a393ccd402bebffaf681b12cea2dcb54"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.718485 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbicana450-account-delete-l2bs5" podStartSLOduration=7.718456237 podStartE2EDuration="7.718456237s" podCreationTimestamp="2025-11-23 07:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:13:47.63310727 +0000 UTC m=+1443.146498573" watchObservedRunningTime="2025-11-23 07:13:47.718456237 +0000 UTC m=+1443.231847540" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.724481 4906 generic.go:334] "Generic (PLEG): container finished" podID="1642db9a-a07d-475a-8112-518d26323d11" containerID="f4c66a184ddf3fddd4f01f4bf5d9927e08a50644685ee93df83ca8a705617d12" exitCode=0 Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.724538 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1642db9a-a07d-475a-8112-518d26323d11","Type":"ContainerDied","Data":"f4c66a184ddf3fddd4f01f4bf5d9927e08a50644685ee93df83ca8a705617d12"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.730713 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron9f60-account-delete-vq457" event={"ID":"00991da6-745e-4757-9d89-a20c5b97e38f","Type":"ContainerStarted","Data":"92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54"} Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.730915 4906 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron9f60-account-delete-vq457" secret="" err="secret \"galera-openstack-dockercfg-gqq6f\" not found" Nov 23 07:13:47 crc kubenswrapper[4906]: I1123 07:13:47.734316 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell082ff-account-delete-s8jc5" podStartSLOduration=6.734295731 podStartE2EDuration="6.734295731s" podCreationTimestamp="2025-11-23 07:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:13:47.712575042 +0000 UTC m=+1443.225966345" watchObservedRunningTime="2025-11-23 07:13:47.734295731 +0000 UTC m=+1443.247687034" Nov 23 07:13:47 crc kubenswrapper[4906]: E1123 07:13:47.817857 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:47 crc kubenswrapper[4906]: E1123 07:13:47.819311 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts podName:00991da6-745e-4757-9d89-a20c5b97e38f nodeName:}" failed. No retries permitted until 2025-11-23 07:13:48.319273948 +0000 UTC m=+1443.832665411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts") pod "neutron9f60-account-delete-vq457" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f") : configmap "openstack-scripts" not found Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.017911 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.039250 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.121560 4906 scope.go:117] "RemoveContainer" containerID="9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.122023 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.123653 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-combined-ca-bundle\") pod \"73149d5c-c16f-4b83-b855-135fc1aab9fe\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.123799 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-config-data\") pod \"73149d5c-c16f-4b83-b855-135fc1aab9fe\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.123836 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73149d5c-c16f-4b83-b855-135fc1aab9fe-logs\") pod \"73149d5c-c16f-4b83-b855-135fc1aab9fe\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.123887 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-internal-tls-certs\") pod \"73149d5c-c16f-4b83-b855-135fc1aab9fe\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.124091 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmwrz\" (UniqueName: \"kubernetes.io/projected/da74c07b-2832-4d8e-9508-415fdc3cac71-kube-api-access-pmwrz\") pod \"da74c07b-2832-4d8e-9508-415fdc3cac71\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.124119 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-public-tls-certs\") pod \"73149d5c-c16f-4b83-b855-135fc1aab9fe\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.124173 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw487\" (UniqueName: \"kubernetes.io/projected/73149d5c-c16f-4b83-b855-135fc1aab9fe-kube-api-access-vw487\") pod \"73149d5c-c16f-4b83-b855-135fc1aab9fe\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.124191 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-config-data\") pod \"da74c07b-2832-4d8e-9508-415fdc3cac71\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.124225 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-scripts\") pod \"73149d5c-c16f-4b83-b855-135fc1aab9fe\" (UID: \"73149d5c-c16f-4b83-b855-135fc1aab9fe\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.124279 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-combined-ca-bundle\") pod \"da74c07b-2832-4d8e-9508-415fdc3cac71\" (UID: \"da74c07b-2832-4d8e-9508-415fdc3cac71\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.130645 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.132201 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73149d5c-c16f-4b83-b855-135fc1aab9fe-logs" (OuterVolumeSpecName: "logs") pod "73149d5c-c16f-4b83-b855-135fc1aab9fe" (UID: "73149d5c-c16f-4b83-b855-135fc1aab9fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.138725 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73149d5c-c16f-4b83-b855-135fc1aab9fe-kube-api-access-vw487" (OuterVolumeSpecName: "kube-api-access-vw487") pod "73149d5c-c16f-4b83-b855-135fc1aab9fe" (UID: "73149d5c-c16f-4b83-b855-135fc1aab9fe"). InnerVolumeSpecName "kube-api-access-vw487". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.142318 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da74c07b-2832-4d8e-9508-415fdc3cac71-kube-api-access-pmwrz" (OuterVolumeSpecName: "kube-api-access-pmwrz") pod "da74c07b-2832-4d8e-9508-415fdc3cac71" (UID: "da74c07b-2832-4d8e-9508-415fdc3cac71"). InnerVolumeSpecName "kube-api-access-pmwrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.146385 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.167133 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-scripts" (OuterVolumeSpecName: "scripts") pod "73149d5c-c16f-4b83-b855-135fc1aab9fe" (UID: "73149d5c-c16f-4b83-b855-135fc1aab9fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.185929 4906 scope.go:117] "RemoveContainer" containerID="209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.192764 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.198917 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da74c07b-2832-4d8e-9508-415fdc3cac71" (UID: "da74c07b-2832-4d8e-9508-415fdc3cac71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.204384 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.237094 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.237830 4906 scope.go:117] "RemoveContainer" containerID="9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.250165 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.252835 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw487\" (UniqueName: \"kubernetes.io/projected/73149d5c-c16f-4b83-b855-135fc1aab9fe-kube-api-access-vw487\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.252955 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.253011 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.253029 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.254192 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73149d5c-c16f-4b83-b855-135fc1aab9fe-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.254341 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.254472 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts podName:01d907ea-d712-40af-b18a-5d55287fbaeb nodeName:}" failed. No retries permitted until 2025-11-23 07:13:49.25445077 +0000 UTC m=+1444.767842083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts") pod "novacell082ff-account-delete-s8jc5" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb") : configmap "openstack-scripts" not found Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.254828 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmwrz\" (UniqueName: \"kubernetes.io/projected/da74c07b-2832-4d8e-9508-415fdc3cac71-kube-api-access-pmwrz\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.255124 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts podName:64c1c989-3443-4b80-9786-56b68660655c nodeName:}" failed. No retries permitted until 2025-11-23 07:13:49.255055977 +0000 UTC m=+1444.768447480 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts") pod "barbicana450-account-delete-l2bs5" (UID: "64c1c989-3443-4b80-9786-56b68660655c") : configmap "openstack-scripts" not found Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.258413 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c\": container with ID starting with 9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c not found: ID does not exist" containerID="9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.258507 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c"} err="failed to get container status \"9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c\": rpc error: code = NotFound desc = could not find container \"9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c\": container with ID starting with 9669a4c83e7bd6bc9667da1f95c15fca8ab462789b8c91ba5932cb76d5f0771c not found: ID does not exist" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.258615 4906 scope.go:117] "RemoveContainer" containerID="209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.260362 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d\": container with ID starting with 209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d not found: ID does not exist" containerID="209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.260422 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d"} err="failed to get container status \"209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d\": rpc error: code = NotFound desc = could not find container \"209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d\": container with ID starting with 209601bb3b002c8b86cf25ee7550abdb08a8f76afaca7b1c14bc382db5fae13d not found: ID does not exist" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.260459 4906 scope.go:117] "RemoveContainer" containerID="8ffeab843681bdc70f3ec1564c88d5f531c4f4543ee3ba842c5877feb9068f85" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.263448 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-config-data" (OuterVolumeSpecName: "config-data") pod "da74c07b-2832-4d8e-9508-415fdc3cac71" (UID: "da74c07b-2832-4d8e-9508-415fdc3cac71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.302717 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-config-data" (OuterVolumeSpecName: "config-data") pod "73149d5c-c16f-4b83-b855-135fc1aab9fe" (UID: "73149d5c-c16f-4b83-b855-135fc1aab9fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.307820 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73149d5c-c16f-4b83-b855-135fc1aab9fe" (UID: "73149d5c-c16f-4b83-b855-135fc1aab9fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.361873 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6g5l\" (UniqueName: \"kubernetes.io/projected/9535028a-afb6-49d6-9bac-541c5324cc55-kube-api-access-p6g5l\") pod \"9535028a-afb6-49d6-9bac-541c5324cc55\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.361946 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-config\") pod \"9535028a-afb6-49d6-9bac-541c5324cc55\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.361986 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-public-tls-certs\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362024 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-logs\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362068 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gch57\" (UniqueName: \"kubernetes.io/projected/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-kube-api-access-gch57\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362110 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362158 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6jg6\" (UniqueName: \"kubernetes.io/projected/1d7b2888-eaed-4879-9f60-e4264aed2287-kube-api-access-v6jg6\") pod \"1d7b2888-eaed-4879-9f60-e4264aed2287\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362181 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-internal-tls-certs\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362220 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-config-data\") pod \"1d7b2888-eaed-4879-9f60-e4264aed2287\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362259 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-scripts\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362277 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-combined-ca-bundle\") pod \"1d7b2888-eaed-4879-9f60-e4264aed2287\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362314 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-internal-tls-certs\") pod \"1d7b2888-eaed-4879-9f60-e4264aed2287\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362339 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-etc-machine-id\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362440 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7b2888-eaed-4879-9f60-e4264aed2287-logs\") pod \"1d7b2888-eaed-4879-9f60-e4264aed2287\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362482 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-combined-ca-bundle\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362522 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-certs\") pod \"9535028a-afb6-49d6-9bac-541c5324cc55\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362543 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-public-tls-certs\") pod \"1d7b2888-eaed-4879-9f60-e4264aed2287\" (UID: \"1d7b2888-eaed-4879-9f60-e4264aed2287\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362573 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-combined-ca-bundle\") pod \"9535028a-afb6-49d6-9bac-541c5324cc55\" (UID: \"9535028a-afb6-49d6-9bac-541c5324cc55\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.362595 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data-custom\") pod \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\" (UID: \"5ee97571-4ae5-4ed0-a32a-67abd5896ae0\") " Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.363501 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da74c07b-2832-4d8e-9508-415fdc3cac71-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.363523 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.363534 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.363625 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.363644 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-logs" (OuterVolumeSpecName: "logs") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.363711 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts podName:00991da6-745e-4757-9d89-a20c5b97e38f nodeName:}" failed. No retries permitted until 2025-11-23 07:13:49.363689913 +0000 UTC m=+1444.877081216 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts") pod "neutron9f60-account-delete-vq457" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f") : configmap "openstack-scripts" not found Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.367984 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-scripts" (OuterVolumeSpecName: "scripts") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.373369 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.374450 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9535028a-afb6-49d6-9bac-541c5324cc55-kube-api-access-p6g5l" (OuterVolumeSpecName: "kube-api-access-p6g5l") pod "9535028a-afb6-49d6-9bac-541c5324cc55" (UID: "9535028a-afb6-49d6-9bac-541c5324cc55"). InnerVolumeSpecName "kube-api-access-p6g5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.375834 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7b2888-eaed-4879-9f60-e4264aed2287-logs" (OuterVolumeSpecName: "logs") pod "1d7b2888-eaed-4879-9f60-e4264aed2287" (UID: "1d7b2888-eaed-4879-9f60-e4264aed2287"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.410610 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7b2888-eaed-4879-9f60-e4264aed2287-kube-api-access-v6jg6" (OuterVolumeSpecName: "kube-api-access-v6jg6") pod "1d7b2888-eaed-4879-9f60-e4264aed2287" (UID: "1d7b2888-eaed-4879-9f60-e4264aed2287"). InnerVolumeSpecName "kube-api-access-v6jg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.415922 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.425465 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-kube-api-access-gch57" (OuterVolumeSpecName: "kube-api-access-gch57") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "kube-api-access-gch57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.476948 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6g5l\" (UniqueName: \"kubernetes.io/projected/9535028a-afb6-49d6-9bac-541c5324cc55-kube-api-access-p6g5l\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.476983 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.476994 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gch57\" (UniqueName: \"kubernetes.io/projected/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-kube-api-access-gch57\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.477004 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6jg6\" (UniqueName: \"kubernetes.io/projected/1d7b2888-eaed-4879-9f60-e4264aed2287-kube-api-access-v6jg6\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.477013 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.477022 4906 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.477029 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7b2888-eaed-4879-9f60-e4264aed2287-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.477038 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.495639 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.539887 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-config-data" (OuterVolumeSpecName: "config-data") pod "1d7b2888-eaed-4879-9f60-e4264aed2287" (UID: "1d7b2888-eaed-4879-9f60-e4264aed2287"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.581985 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.584380 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.604384 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "9535028a-afb6-49d6-9bac-541c5324cc55" (UID: "9535028a-afb6-49d6-9bac-541c5324cc55"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.619890 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9535028a-afb6-49d6-9bac-541c5324cc55" (UID: "9535028a-afb6-49d6-9bac-541c5324cc55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.628878 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "73149d5c-c16f-4b83-b855-135fc1aab9fe" (UID: "73149d5c-c16f-4b83-b855-135fc1aab9fe"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.647110 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1d7b2888-eaed-4879-9f60-e4264aed2287" (UID: "1d7b2888-eaed-4879-9f60-e4264aed2287"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.680535 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "9535028a-afb6-49d6-9bac-541c5324cc55" (UID: "9535028a-afb6-49d6-9bac-541c5324cc55"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.686809 4906 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.686849 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.686863 4906 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9535028a-afb6-49d6-9bac-541c5324cc55-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.686874 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.686886 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.712665 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d7b2888-eaed-4879-9f60-e4264aed2287" (UID: "1d7b2888-eaed-4879-9f60-e4264aed2287"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.724287 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.726178 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.735125 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "73149d5c-c16f-4b83-b855-135fc1aab9fe" (UID: "73149d5c-c16f-4b83-b855-135fc1aab9fe"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.744946 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.751887 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.745034 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="ovn-northd" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.765818 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1d7b2888-eaed-4879-9f60-e4264aed2287" (UID: "1d7b2888-eaed-4879-9f60-e4264aed2287"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.778044 4906 generic.go:334] "Generic (PLEG): container finished" podID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerID="f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df" exitCode=0 Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.778288 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerDied","Data":"f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.779931 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder46ad-account-delete-rpn4p" event={"ID":"4df634a0-630f-46a5-b304-86211157136a","Type":"ContainerDied","Data":"55bcf6a4b1677ea5ec89ebefc19f137bea7a0453cd26cf7bf4cc78e4e7d235ec"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.780058 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55bcf6a4b1677ea5ec89ebefc19f137bea7a0453cd26cf7bf4cc78e4e7d235ec" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.782467 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.782665 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1d7b2888-eaed-4879-9f60-e4264aed2287","Type":"ContainerDied","Data":"44520f7b636e7cc8da583f7ba7d9f3522a3d30b934dfeafc72f7c6f3464440a6"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.789181 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73149d5c-c16f-4b83-b855-135fc1aab9fe-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.789214 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.789226 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7b2888-eaed-4879-9f60-e4264aed2287-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.789237 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.789331 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:48 crc kubenswrapper[4906]: E1123 07:13:48.789383 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data podName:4d677a57-1acf-4627-aa66-f4bee96a2b51 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:56.789367486 +0000 UTC m=+1452.302758789 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data") pod "rabbitmq-cell1-server-0" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51") : configmap "rabbitmq-cell1-config-data" not found Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.790533 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bdb7fc4cb-h962s" event={"ID":"4aedece6-346c-4003-8d00-95509a35f23b","Type":"ContainerDied","Data":"84b441d187fa8cdc082ceec636414ade0519a551c14dc2185e814ec8ae3a3647"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.790621 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84b441d187fa8cdc082ceec636414ade0519a551c14dc2185e814ec8ae3a3647" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.792651 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.794877 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9535028a-afb6-49d6-9bac-541c5324cc55","Type":"ContainerDied","Data":"11579df96862d0b66237165e68aa09a4c23bfea76d77ad2076e3e62b0c745db8"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.794964 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.809212 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00d152f2-1543-4b14-b214-73747a3f9a26","Type":"ContainerDied","Data":"c1ba7e6a6a3f0bc8878932b83d10ea4c8acb3eabbbf1c12aed17e0000d862edf"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.809241 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1ba7e6a6a3f0bc8878932b83d10ea4c8acb3eabbbf1c12aed17e0000d862edf" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.810938 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data" (OuterVolumeSpecName: "config-data") pod "5ee97571-4ae5-4ed0-a32a-67abd5896ae0" (UID: "5ee97571-4ae5-4ed0-a32a-67abd5896ae0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.811870 4906 generic.go:334] "Generic (PLEG): container finished" podID="27d994c3-4d19-4084-8c09-d0f2a8d7ede7" containerID="c62d53e7d032049f41cdae0eac7ed993f6453996187bc0ab5a5c3458db844d22" exitCode=0 Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.811916 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"27d994c3-4d19-4084-8c09-d0f2a8d7ede7","Type":"ContainerDied","Data":"c62d53e7d032049f41cdae0eac7ed993f6453996187bc0ab5a5c3458db844d22"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.814263 4906 generic.go:334] "Generic (PLEG): container finished" podID="212ee22b-c3e7-4d9c-b952-5837b5982c63" containerID="4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef" exitCode=0 Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.814320 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"212ee22b-c3e7-4d9c-b952-5837b5982c63","Type":"ContainerDied","Data":"4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.816481 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1642db9a-a07d-475a-8112-518d26323d11","Type":"ContainerDied","Data":"116a985e4f0a6ad1024c63bfe365d28dafd484ecb5fc5da136db31844fa21b2b"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.816505 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="116a985e4f0a6ad1024c63bfe365d28dafd484ecb5fc5da136db31844fa21b2b" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.833412 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d","Type":"ContainerDied","Data":"da03f1fec6165233dbdd91e71071a0ec9a3f6fb4ff76060cfa558c2ecdcd499d"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.833456 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da03f1fec6165233dbdd91e71071a0ec9a3f6fb4ff76060cfa558c2ecdcd499d" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.837917 4906 generic.go:334] "Generic (PLEG): container finished" podID="1ab3c4dd-5b70-4127-aa8d-570ffcde7477" containerID="a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33" exitCode=0 Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.837981 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1ab3c4dd-5b70-4127-aa8d-570ffcde7477","Type":"ContainerDied","Data":"a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.840844 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementb64f-account-delete-576n6" event={"ID":"6093f84a-c555-4fc4-a207-f54e5ddf2a85","Type":"ContainerDied","Data":"e161829b4c3412d3950ffec321774284108121997a6f8f59cc52bdcd36b4f608"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.840870 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e161829b4c3412d3950ffec321774284108121997a6f8f59cc52bdcd36b4f608" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.844898 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ee97571-4ae5-4ed0-a32a-67abd5896ae0","Type":"ContainerDied","Data":"b8acffca0291884aad6ad380d3e1bceec373abb266c89d76eeec84689419d1b3"} Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.844992 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.852113 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5d7f54fb65-944h6" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.193:5353: i/o timeout" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.857661 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ffb7cbc86-lz7bc" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.857876 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.872175 4906 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicana450-account-delete-l2bs5" secret="" err="secret \"galera-openstack-dockercfg-gqq6f\" not found" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.872638 4906 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell082ff-account-delete-s8jc5" secret="" err="secret \"galera-openstack-dockercfg-gqq6f\" not found" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.872779 4906 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron9f60-account-delete-vq457" secret="" err="secret \"galera-openstack-dockercfg-gqq6f\" not found" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.892818 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.892848 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ee97571-4ae5-4ed0-a32a-67abd5896ae0-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:48 crc kubenswrapper[4906]: I1123 07:13:48.976787 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": dial tcp 10.217.0.196:3000: connect: connection refused" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.120399 4906 scope.go:117] "RemoveContainer" containerID="ed942ae5cec5564c89129803c0336b09e9319da12fa2443229288d6c827463c8" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.121969 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.138726 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.156101 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.181663 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.191949 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.196658 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.198387 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4df634a0-630f-46a5-b304-86211157136a-operator-scripts\") pod \"4df634a0-630f-46a5-b304-86211157136a\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.199386 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6093f84a-c555-4fc4-a207-f54e5ddf2a85-operator-scripts\") pod \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.199660 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlvsk\" (UniqueName: \"kubernetes.io/projected/1642db9a-a07d-475a-8112-518d26323d11-kube-api-access-hlvsk\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.199777 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-httpd-run\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.199864 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-scripts\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.199971 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-logs\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.200270 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-config-data\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.200373 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvxnr\" (UniqueName: \"kubernetes.io/projected/4df634a0-630f-46a5-b304-86211157136a-kube-api-access-hvxnr\") pod \"4df634a0-630f-46a5-b304-86211157136a\" (UID: \"4df634a0-630f-46a5-b304-86211157136a\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.200555 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.200718 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-internal-tls-certs\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.200851 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmvjr\" (UniqueName: \"kubernetes.io/projected/6093f84a-c555-4fc4-a207-f54e5ddf2a85-kube-api-access-tmvjr\") pod \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\" (UID: \"6093f84a-c555-4fc4-a207-f54e5ddf2a85\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.200939 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-combined-ca-bundle\") pod \"1642db9a-a07d-475a-8112-518d26323d11\" (UID: \"1642db9a-a07d-475a-8112-518d26323d11\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.204091 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.198891 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4df634a0-630f-46a5-b304-86211157136a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4df634a0-630f-46a5-b304-86211157136a" (UID: "4df634a0-630f-46a5-b304-86211157136a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.204911 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-logs" (OuterVolumeSpecName: "logs") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.205252 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.205635 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6093f84a-c555-4fc4-a207-f54e5ddf2a85-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6093f84a-c555-4fc4-a207-f54e5ddf2a85" (UID: "6093f84a-c555-4fc4-a207-f54e5ddf2a85"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.209850 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.220582 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1642db9a-a07d-475a-8112-518d26323d11-kube-api-access-hlvsk" (OuterVolumeSpecName: "kube-api-access-hlvsk") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "kube-api-access-hlvsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.239898 4906 scope.go:117] "RemoveContainer" containerID="a091f675900348c0b86ab0b4409f0e26e962032b4b02dd301de38636bc1e7e50" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.252999 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.256483 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.260955 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-scripts" (OuterVolumeSpecName: "scripts") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.289288 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.305040 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6093f84a-c555-4fc4-a207-f54e5ddf2a85-kube-api-access-tmvjr" (OuterVolumeSpecName: "kube-api-access-tmvjr") pod "6093f84a-c555-4fc4-a207-f54e5ddf2a85" (UID: "6093f84a-c555-4fc4-a207-f54e5ddf2a85"). InnerVolumeSpecName "kube-api-access-tmvjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.305882 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df634a0-630f-46a5-b304-86211157136a-kube-api-access-hvxnr" (OuterVolumeSpecName: "kube-api-access-hvxnr") pod "4df634a0-630f-46a5-b304-86211157136a" (UID: "4df634a0-630f-46a5-b304-86211157136a"). InnerVolumeSpecName "kube-api-access-hvxnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.325056 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.339471 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data\") pod \"4aedece6-346c-4003-8d00-95509a35f23b\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340031 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-public-tls-certs\") pod \"4aedece6-346c-4003-8d00-95509a35f23b\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340145 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-nova-metadata-tls-certs\") pod \"00d152f2-1543-4b14-b214-73747a3f9a26\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340240 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w7n6\" (UniqueName: \"kubernetes.io/projected/4aedece6-346c-4003-8d00-95509a35f23b-kube-api-access-9w7n6\") pod \"4aedece6-346c-4003-8d00-95509a35f23b\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340377 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data-custom\") pod \"4aedece6-346c-4003-8d00-95509a35f23b\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340454 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-combined-ca-bundle\") pod \"00d152f2-1543-4b14-b214-73747a3f9a26\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340649 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqpwt\" (UniqueName: \"kubernetes.io/projected/00d152f2-1543-4b14-b214-73747a3f9a26-kube-api-access-xqpwt\") pod \"00d152f2-1543-4b14-b214-73747a3f9a26\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340884 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-internal-tls-certs\") pod \"4aedece6-346c-4003-8d00-95509a35f23b\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.340974 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-combined-ca-bundle\") pod \"4aedece6-346c-4003-8d00-95509a35f23b\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.341041 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00d152f2-1543-4b14-b214-73747a3f9a26-logs\") pod \"00d152f2-1543-4b14-b214-73747a3f9a26\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.341143 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aedece6-346c-4003-8d00-95509a35f23b-logs\") pod \"4aedece6-346c-4003-8d00-95509a35f23b\" (UID: \"4aedece6-346c-4003-8d00-95509a35f23b\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.342452 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-config-data\") pod \"00d152f2-1543-4b14-b214-73747a3f9a26\" (UID: \"00d152f2-1543-4b14-b214-73747a3f9a26\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.346597 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d152f2-1543-4b14-b214-73747a3f9a26-logs" (OuterVolumeSpecName: "logs") pod "00d152f2-1543-4b14-b214-73747a3f9a26" (UID: "00d152f2-1543-4b14-b214-73747a3f9a26"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.347585 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aedece6-346c-4003-8d00-95509a35f23b-logs" (OuterVolumeSpecName: "logs") pod "4aedece6-346c-4003-8d00-95509a35f23b" (UID: "4aedece6-346c-4003-8d00-95509a35f23b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.347736 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.347822 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts podName:64c1c989-3443-4b80-9786-56b68660655c nodeName:}" failed. No retries permitted until 2025-11-23 07:13:51.347801659 +0000 UTC m=+1446.861192962 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts") pod "barbicana450-account-delete-l2bs5" (UID: "64c1c989-3443-4b80-9786-56b68660655c") : configmap "openstack-scripts" not found Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.347926 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.347963 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts podName:01d907ea-d712-40af-b18a-5d55287fbaeb nodeName:}" failed. No retries permitted until 2025-11-23 07:13:51.347954883 +0000 UTC m=+1446.861346186 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts") pod "novacell082ff-account-delete-s8jc5" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb") : configmap "openstack-scripts" not found Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348080 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00d152f2-1543-4b14-b214-73747a3f9a26-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348099 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlvsk\" (UniqueName: \"kubernetes.io/projected/1642db9a-a07d-475a-8112-518d26323d11-kube-api-access-hlvsk\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348116 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aedece6-346c-4003-8d00-95509a35f23b-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348127 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348142 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348155 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1642db9a-a07d-475a-8112-518d26323d11-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348166 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvxnr\" (UniqueName: \"kubernetes.io/projected/4df634a0-630f-46a5-b304-86211157136a-kube-api-access-hvxnr\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348199 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348216 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348229 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmvjr\" (UniqueName: \"kubernetes.io/projected/6093f84a-c555-4fc4-a207-f54e5ddf2a85-kube-api-access-tmvjr\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348242 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348257 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4df634a0-630f-46a5-b304-86211157136a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.348274 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6093f84a-c555-4fc4-a207-f54e5ddf2a85-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.390535 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d152f2-1543-4b14-b214-73747a3f9a26-kube-api-access-xqpwt" (OuterVolumeSpecName: "kube-api-access-xqpwt") pod "00d152f2-1543-4b14-b214-73747a3f9a26" (UID: "00d152f2-1543-4b14-b214-73747a3f9a26"). InnerVolumeSpecName "kube-api-access-xqpwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.397197 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4aedece6-346c-4003-8d00-95509a35f23b" (UID: "4aedece6-346c-4003-8d00-95509a35f23b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.399197 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aedece6-346c-4003-8d00-95509a35f23b-kube-api-access-9w7n6" (OuterVolumeSpecName: "kube-api-access-9w7n6") pod "4aedece6-346c-4003-8d00-95509a35f23b" (UID: "4aedece6-346c-4003-8d00-95509a35f23b"). InnerVolumeSpecName "kube-api-access-9w7n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.402348 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" path="/var/lib/kubelet/pods/1d7b2888-eaed-4879-9f60-e4264aed2287/volumes" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.403124 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40cb3343-177c-42cf-8604-c22eaca5e593" path="/var/lib/kubelet/pods/40cb3343-177c-42cf-8604-c22eaca5e593/volumes" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.403870 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" path="/var/lib/kubelet/pods/51dd433b-f1a2-422a-bf5b-643e150c3bca/volumes" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.405495 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" path="/var/lib/kubelet/pods/5ee97571-4ae5-4ed0-a32a-67abd5896ae0/volumes" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.407401 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cee328a-b746-486d-850b-148ca1e66eea" path="/var/lib/kubelet/pods/6cee328a-b746-486d-850b-148ca1e66eea/volumes" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.408034 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78dd4ca-5705-4554-bdc0-1ec212b5751e" path="/var/lib/kubelet/pods/e78dd4ca-5705-4554-bdc0-1ec212b5751e/volumes" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.451147 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w7n6\" (UniqueName: \"kubernetes.io/projected/4aedece6-346c-4003-8d00-95509a35f23b-kube-api-access-9w7n6\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.453878 4906 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.453983 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data podName:9a0222ea-7767-4d08-a6ec-6659a33f9df2 nodeName:}" failed. No retries permitted until 2025-11-23 07:13:57.453954861 +0000 UTC m=+1452.967346164 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data") pod "rabbitmq-server-0" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2") : configmap "rabbitmq-config-data" not found Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.454154 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:49 crc kubenswrapper[4906]: E1123 07:13:49.454248 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts podName:00991da6-745e-4757-9d89-a20c5b97e38f nodeName:}" failed. No retries permitted until 2025-11-23 07:13:51.454224758 +0000 UTC m=+1446.967616061 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts") pod "neutron9f60-account-delete-vq457" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f") : configmap "openstack-scripts" not found Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.454520 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.454738 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqpwt\" (UniqueName: \"kubernetes.io/projected/00d152f2-1543-4b14-b214-73747a3f9a26-kube-api-access-xqpwt\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.468897 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-config-data" (OuterVolumeSpecName: "config-data") pod "1642db9a-a07d-475a-8112-518d26323d11" (UID: "1642db9a-a07d-475a-8112-518d26323d11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.493871 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-config-data" (OuterVolumeSpecName: "config-data") pod "00d152f2-1543-4b14-b214-73747a3f9a26" (UID: "00d152f2-1543-4b14-b214-73747a3f9a26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.499235 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.504245 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4aedece6-346c-4003-8d00-95509a35f23b" (UID: "4aedece6-346c-4003-8d00-95509a35f23b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.521931 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.521979 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.521998 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.522011 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.522022 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ffb7cbc86-lz7bc"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.522037 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ffb7cbc86-lz7bc"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.523418 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.531489 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data" (OuterVolumeSpecName: "config-data") pod "4aedece6-346c-4003-8d00-95509a35f23b" (UID: "4aedece6-346c-4003-8d00-95509a35f23b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.538190 4906 scope.go:117] "RemoveContainer" containerID="2973c1fcb92c6b4358da58cb882d94a8ee6b77fc0c57c5eca2bbf4845fe195a0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.554519 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4aedece6-346c-4003-8d00-95509a35f23b" (UID: "4aedece6-346c-4003-8d00-95509a35f23b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.559324 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.561402 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.561892 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.564912 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.564930 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.564942 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.564955 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.564964 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1642db9a-a07d-475a-8112-518d26323d11-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.595205 4906 scope.go:117] "RemoveContainer" containerID="77ce16e41b99ad023af4aa1620f90f3b6cbe24953a476cead47e6c790b48478b" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.597307 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aedece6-346c-4003-8d00-95509a35f23b" (UID: "4aedece6-346c-4003-8d00-95509a35f23b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.608288 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "00d152f2-1543-4b14-b214-73747a3f9a26" (UID: "00d152f2-1543-4b14-b214-73747a3f9a26"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.624146 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.631402 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00d152f2-1543-4b14-b214-73747a3f9a26" (UID: "00d152f2-1543-4b14-b214-73747a3f9a26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.651437 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.654239 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.665981 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-config-data\") pod \"212ee22b-c3e7-4d9c-b952-5837b5982c63\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666053 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666138 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfh95\" (UniqueName: \"kubernetes.io/projected/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-kube-api-access-lfh95\") pod \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666184 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-config-data\") pod \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666206 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-combined-ca-bundle\") pod \"212ee22b-c3e7-4d9c-b952-5837b5982c63\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666262 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-public-tls-certs\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666313 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h89fv\" (UniqueName: \"kubernetes.io/projected/212ee22b-c3e7-4d9c-b952-5837b5982c63-kube-api-access-h89fv\") pod \"212ee22b-c3e7-4d9c-b952-5837b5982c63\" (UID: \"212ee22b-c3e7-4d9c-b952-5837b5982c63\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666346 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckv26\" (UniqueName: \"kubernetes.io/projected/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-kube-api-access-ckv26\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666371 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-config-data\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666429 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-scripts\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666454 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-combined-ca-bundle\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666543 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-logs\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.666570 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-httpd-run\") pod \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\" (UID: \"a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.667191 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-combined-ca-bundle\") pod \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\" (UID: \"1ab3c4dd-5b70-4127-aa8d-570ffcde7477\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.668729 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aedece6-346c-4003-8d00-95509a35f23b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.668747 4906 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.668803 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d152f2-1543-4b14-b214-73747a3f9a26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.672371 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-logs" (OuterVolumeSpecName: "logs") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.672738 4906 scope.go:117] "RemoveContainer" containerID="d7113d47184e50909b0d719000d22c051debd5b0dca75d3f957a7612dc3670a1" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.674100 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.681207 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-scripts" (OuterVolumeSpecName: "scripts") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.681405 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-kube-api-access-lfh95" (OuterVolumeSpecName: "kube-api-access-lfh95") pod "1ab3c4dd-5b70-4127-aa8d-570ffcde7477" (UID: "1ab3c4dd-5b70-4127-aa8d-570ffcde7477"). InnerVolumeSpecName "kube-api-access-lfh95". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.682237 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/212ee22b-c3e7-4d9c-b952-5837b5982c63-kube-api-access-h89fv" (OuterVolumeSpecName: "kube-api-access-h89fv") pod "212ee22b-c3e7-4d9c-b952-5837b5982c63" (UID: "212ee22b-c3e7-4d9c-b952-5837b5982c63"). InnerVolumeSpecName "kube-api-access-h89fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.684841 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.689895 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-kube-api-access-ckv26" (OuterVolumeSpecName: "kube-api-access-ckv26") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "kube-api-access-ckv26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.709207 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-config-data" (OuterVolumeSpecName: "config-data") pod "1ab3c4dd-5b70-4127-aa8d-570ffcde7477" (UID: "1ab3c4dd-5b70-4127-aa8d-570ffcde7477"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.721011 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-config-data" (OuterVolumeSpecName: "config-data") pod "212ee22b-c3e7-4d9c-b952-5837b5982c63" (UID: "212ee22b-c3e7-4d9c-b952-5837b5982c63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.721068 4906 scope.go:117] "RemoveContainer" containerID="0ef240d12f0659c3ea6a9b32c942f26f9370dad6938a45b1af45fb31d908312f" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.733806 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ab3c4dd-5b70-4127-aa8d-570ffcde7477" (UID: "1ab3c4dd-5b70-4127-aa8d-570ffcde7477"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.737775 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "212ee22b-c3e7-4d9c-b952-5837b5982c63" (UID: "212ee22b-c3e7-4d9c-b952-5837b5982c63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.754052 4906 scope.go:117] "RemoveContainer" containerID="e82535f708e61c7ee739e897ea48ec09a393ccd402bebffaf681b12cea2dcb54" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.755311 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.756399 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772400 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab42792-22e3-48ee-8028-3d181679190c-operator-scripts\") pod \"0ab42792-22e3-48ee-8028-3d181679190c\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772584 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-config-data\") pod \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772637 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkbkq\" (UniqueName: \"kubernetes.io/projected/0ab42792-22e3-48ee-8028-3d181679190c-kube-api-access-vkbkq\") pod \"0ab42792-22e3-48ee-8028-3d181679190c\" (UID: \"0ab42792-22e3-48ee-8028-3d181679190c\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772734 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-combined-ca-bundle\") pod \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772784 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p9gc\" (UniqueName: \"kubernetes.io/projected/2e0ab9b5-09c4-49b9-a112-76c18f06524d-kube-api-access-9p9gc\") pod \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772845 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-memcached-tls-certs\") pod \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772928 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2928c\" (UniqueName: \"kubernetes.io/projected/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kube-api-access-2928c\") pod \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772967 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e0ab9b5-09c4-49b9-a112-76c18f06524d-operator-scripts\") pod \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\" (UID: \"2e0ab9b5-09c4-49b9-a112-76c18f06524d\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.772990 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kolla-config\") pod \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\" (UID: \"27d994c3-4d19-4084-8c09-d0f2a8d7ede7\") " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.773352 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ab42792-22e3-48ee-8028-3d181679190c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0ab42792-22e3-48ee-8028-3d181679190c" (UID: "0ab42792-22e3-48ee-8028-3d181679190c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.773379 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-config-data" (OuterVolumeSpecName: "config-data") pod "27d994c3-4d19-4084-8c09-d0f2a8d7ede7" (UID: "27d994c3-4d19-4084-8c09-d0f2a8d7ede7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.773667 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776043 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab42792-22e3-48ee-8028-3d181679190c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776073 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776086 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfh95\" (UniqueName: \"kubernetes.io/projected/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-kube-api-access-lfh95\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776099 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776109 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/212ee22b-c3e7-4d9c-b952-5837b5982c63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776118 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776128 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776138 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h89fv\" (UniqueName: \"kubernetes.io/projected/212ee22b-c3e7-4d9c-b952-5837b5982c63-kube-api-access-h89fv\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776149 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckv26\" (UniqueName: \"kubernetes.io/projected/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-kube-api-access-ckv26\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776159 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776174 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776185 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776197 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776206 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ab3c4dd-5b70-4127-aa8d-570ffcde7477-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.774876 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "27d994c3-4d19-4084-8c09-d0f2a8d7ede7" (UID: "27d994c3-4d19-4084-8c09-d0f2a8d7ede7"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.775234 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-config-data" (OuterVolumeSpecName: "config-data") pod "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" (UID: "a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.775859 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e0ab9b5-09c4-49b9-a112-76c18f06524d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e0ab9b5-09c4-49b9-a112-76c18f06524d" (UID: "2e0ab9b5-09c4-49b9-a112-76c18f06524d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.776578 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e0ab9b5-09c4-49b9-a112-76c18f06524d-kube-api-access-9p9gc" (OuterVolumeSpecName: "kube-api-access-9p9gc") pod "2e0ab9b5-09c4-49b9-a112-76c18f06524d" (UID: "2e0ab9b5-09c4-49b9-a112-76c18f06524d"). InnerVolumeSpecName "kube-api-access-9p9gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.777345 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ab42792-22e3-48ee-8028-3d181679190c-kube-api-access-vkbkq" (OuterVolumeSpecName: "kube-api-access-vkbkq") pod "0ab42792-22e3-48ee-8028-3d181679190c" (UID: "0ab42792-22e3-48ee-8028-3d181679190c"). InnerVolumeSpecName "kube-api-access-vkbkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.779081 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kube-api-access-2928c" (OuterVolumeSpecName: "kube-api-access-2928c") pod "27d994c3-4d19-4084-8c09-d0f2a8d7ede7" (UID: "27d994c3-4d19-4084-8c09-d0f2a8d7ede7"). InnerVolumeSpecName "kube-api-access-2928c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.804557 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.817802 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27d994c3-4d19-4084-8c09-d0f2a8d7ede7" (UID: "27d994c3-4d19-4084-8c09-d0f2a8d7ede7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.835276 4906 scope.go:117] "RemoveContainer" containerID="df25a67d5a9f5925815eee2060a4fdcc509bfcf08b8d1d1a7f6827a40492c808" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.850660 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "27d994c3-4d19-4084-8c09-d0f2a8d7ede7" (UID: "27d994c3-4d19-4084-8c09-d0f2a8d7ede7"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.875494 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi944e-account-delete-gw8gk" event={"ID":"0ab42792-22e3-48ee-8028-3d181679190c","Type":"ContainerDied","Data":"d56f185ae462d261cefb6e5dbaf0ecb893e6dedd267291f9b4fa4cdb06b06e4b"} Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.875547 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d56f185ae462d261cefb6e5dbaf0ecb893e6dedd267291f9b4fa4cdb06b06e4b" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.875611 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi944e-account-delete-gw8gk" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.878943 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2928c\" (UniqueName: \"kubernetes.io/projected/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kube-api-access-2928c\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.878971 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e0ab9b5-09c4-49b9-a112-76c18f06524d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.878982 4906 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.878992 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.879002 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkbkq\" (UniqueName: \"kubernetes.io/projected/0ab42792-22e3-48ee-8028-3d181679190c-kube-api-access-vkbkq\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.879011 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.879019 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.879027 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p9gc\" (UniqueName: \"kubernetes.io/projected/2e0ab9b5-09c4-49b9-a112-76c18f06524d-kube-api-access-9p9gc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.879038 4906 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d994c3-4d19-4084-8c09-d0f2a8d7ede7-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.882291 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"212ee22b-c3e7-4d9c-b952-5837b5982c63","Type":"ContainerDied","Data":"7f37ec9bea7956b87da8f1438c4bc7e51543df6e6a3ff0f6230dcb171cf72e7f"} Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.882413 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.905555 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.905562 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1ab3c4dd-5b70-4127-aa8d-570ffcde7477","Type":"ContainerDied","Data":"9649c64a7d519f1db26ea6b55d1a041e210285132102870eff86164c7ebc9e8a"} Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.915848 4906 scope.go:117] "RemoveContainer" containerID="4b63433577603fda2887f997c3f7ed29bb055097047f7bb248652f7fa8e176ef" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.917313 4906 generic.go:334] "Generic (PLEG): container finished" podID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerID="d83546ce8058eec4db9f5c5b6441b7ce247b7a874eb9b2018665ca353691af54" exitCode=0 Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.917873 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d677a57-1acf-4627-aa66-f4bee96a2b51","Type":"ContainerDied","Data":"d83546ce8058eec4db9f5c5b6441b7ce247b7a874eb9b2018665ca353691af54"} Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.937735 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi944e-account-delete-gw8gk"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.943940 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceaaf0-account-delete-j59k7" event={"ID":"2e0ab9b5-09c4-49b9-a112-76c18f06524d","Type":"ContainerDied","Data":"d0f6fec4ee2f6b122dca6c37a54aed3f3e5cc1be47767f587d23965550258531"} Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.944109 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0f6fec4ee2f6b122dca6c37a54aed3f3e5cc1be47767f587d23965550258531" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.944121 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanceaaf0-account-delete-j59k7" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.956933 4906 generic.go:334] "Generic (PLEG): container finished" podID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerID="27497e40b508a21677f03a6794d604b1ad821f2140f70be90b3a8b007d2d719d" exitCode=0 Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.957042 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adbcc53a-3e0f-47b0-a028-01c686a18205","Type":"ContainerDied","Data":"27497e40b508a21677f03a6794d604b1ad821f2140f70be90b3a8b007d2d719d"} Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.963839 4906 scope.go:117] "RemoveContainer" containerID="a65190cba7a8bf0748129a202b35977859662b95dc053d39464daaab84ffcb33" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.969183 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi944e-account-delete-gw8gk"] Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.969935 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementb64f-account-delete-576n6" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.970138 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.970526 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.970840 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bdb7fc4cb-h962s" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.971340 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.971355 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"27d994c3-4d19-4084-8c09-d0f2a8d7ede7","Type":"ContainerDied","Data":"8527db811030ccea75981691598f1f7afdfe79a7d8eb1407e57d34e3f5276308"} Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.971764 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.971805 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder46ad-account-delete-rpn4p" Nov 23 07:13:49 crc kubenswrapper[4906]: I1123 07:13:49.992019 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.000689 4906 scope.go:117] "RemoveContainer" containerID="c62d53e7d032049f41cdae0eac7ed993f6453996187bc0ab5a5c3458db844d22" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.022622 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.032695 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.041939 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.052632 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.071807 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.082495 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.088604 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.095395 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.098119 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.101862 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.108520 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bdb7fc4cb-h962s"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.113383 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-bdb7fc4cb-h962s"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.126503 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.133784 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183357 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-generated\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183412 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-galera-tls-certs\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183495 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-operator-scripts\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183515 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-default\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183593 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prddx\" (UniqueName: \"kubernetes.io/projected/adbcc53a-3e0f-47b0-a028-01c686a18205-kube-api-access-prddx\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183627 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-combined-ca-bundle\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183737 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-kolla-config\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.183922 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"adbcc53a-3e0f-47b0-a028-01c686a18205\" (UID: \"adbcc53a-3e0f-47b0-a028-01c686a18205\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.185001 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.185924 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.186323 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.189664 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.192969 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adbcc53a-3e0f-47b0-a028-01c686a18205-kube-api-access-prddx" (OuterVolumeSpecName: "kube-api-access-prddx") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "kube-api-access-prddx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.213189 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.233661 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.262354 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "adbcc53a-3e0f-47b0-a028-01c686a18205" (UID: "adbcc53a-3e0f-47b0-a028-01c686a18205"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288485 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288775 4906 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288809 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288820 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288831 4906 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/adbcc53a-3e0f-47b0-a028-01c686a18205-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288839 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288851 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/adbcc53a-3e0f-47b0-a028-01c686a18205-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.288859 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prddx\" (UniqueName: \"kubernetes.io/projected/adbcc53a-3e0f-47b0-a028-01c686a18205-kube-api-access-prddx\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.323559 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.397993 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.439023 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.601704 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-server-conf\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.601797 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-plugins\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.601826 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-plugins-conf\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.601867 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrvvj\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-kube-api-access-xrvvj\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.601934 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-tls\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.602002 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d677a57-1acf-4627-aa66-f4bee96a2b51-pod-info\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.602030 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d677a57-1acf-4627-aa66-f4bee96a2b51-erlang-cookie-secret\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.602885 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-erlang-cookie\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.602930 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-confd\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.602975 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.603014 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"4d677a57-1acf-4627-aa66-f4bee96a2b51\" (UID: \"4d677a57-1acf-4627-aa66-f4bee96a2b51\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.613032 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.614471 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.614896 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.617310 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.622641 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.624199 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4d677a57-1acf-4627-aa66-f4bee96a2b51-pod-info" (OuterVolumeSpecName: "pod-info") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.629851 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d677a57-1acf-4627-aa66-f4bee96a2b51-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.635402 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data" (OuterVolumeSpecName: "config-data") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.636120 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-kube-api-access-xrvvj" (OuterVolumeSpecName: "kube-api-access-xrvvj") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "kube-api-access-xrvvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.696215 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-server-conf" (OuterVolumeSpecName: "server-conf") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705488 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705526 4906 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4d677a57-1acf-4627-aa66-f4bee96a2b51-pod-info\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705536 4906 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4d677a57-1acf-4627-aa66-f4bee96a2b51-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705547 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705558 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705591 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705600 4906 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-server-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705610 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705618 4906 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4d677a57-1acf-4627-aa66-f4bee96a2b51-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.705627 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrvvj\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-kube-api-access-xrvvj\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.726502 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.732017 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="6cee328a-b746-486d-850b-148ca1e66eea" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.192:6080/vnc_lite.html\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.751601 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4d677a57-1acf-4627-aa66-f4bee96a2b51" (UID: "4d677a57-1acf-4627-aa66-f4bee96a2b51"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.759707 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.772932 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.810063 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4d677a57-1acf-4627-aa66-f4bee96a2b51-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.810099 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911554 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-plugins-conf\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911702 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-public-tls-certs\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911793 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-tls\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911826 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-scripts\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911859 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a0222ea-7767-4d08-a6ec-6659a33f9df2-pod-info\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911881 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vpbf\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-kube-api-access-7vpbf\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911945 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-confd\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.911985 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-internal-tls-certs\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.912047 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-credential-keys\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.912079 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9gbw\" (UniqueName: \"kubernetes.io/projected/2e303c56-48b8-4176-99f6-fd0aeb26cf94-kube-api-access-d9gbw\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.912131 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-combined-ca-bundle\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.912163 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-erlang-cookie\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.913250 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.915744 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.916211 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.917153 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-plugins\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.917210 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-config-data\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.917252 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-server-conf\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.917282 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.917305 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a0222ea-7767-4d08-a6ec-6659a33f9df2-erlang-cookie-secret\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.917323 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-fernet-keys\") pod \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\" (UID: \"2e303c56-48b8-4176-99f6-fd0aeb26cf94\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.917344 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\" (UID: \"9a0222ea-7767-4d08-a6ec-6659a33f9df2\") " Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.918037 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.918078 4906 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.918090 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.918953 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.930128 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-scripts" (OuterVolumeSpecName: "scripts") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.931529 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-kube-api-access-7vpbf" (OuterVolumeSpecName: "kube-api-access-7vpbf") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "kube-api-access-7vpbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.943724 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.947138 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.947199 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.954506 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-vzfd7"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.971124 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.971143 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9a0222ea-7767-4d08-a6ec-6659a33f9df2-pod-info" (OuterVolumeSpecName: "pod-info") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.971219 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e303c56-48b8-4176-99f6-fd0aeb26cf94-kube-api-access-d9gbw" (OuterVolumeSpecName: "kube-api-access-d9gbw") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "kube-api-access-d9gbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.971345 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.972365 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-vzfd7"] Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.978361 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0222ea-7767-4d08-a6ec-6659a33f9df2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.983388 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_63999fb5-88e7-4df1-8084-267d0e37ac4c/ovn-northd/0.log" Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.983436 4906 generic.go:334] "Generic (PLEG): container finished" podID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" exitCode=139 Nov 23 07:13:50 crc kubenswrapper[4906]: I1123 07:13:50.983497 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63999fb5-88e7-4df1-8084-267d0e37ac4c","Type":"ContainerDied","Data":"e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65"} Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.005866 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a450-account-create-8nn5h"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.007379 4906 generic.go:334] "Generic (PLEG): container finished" podID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerID="bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af" exitCode=0 Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.007479 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a0222ea-7767-4d08-a6ec-6659a33f9df2","Type":"ContainerDied","Data":"bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af"} Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.007554 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9a0222ea-7767-4d08-a6ec-6659a33f9df2","Type":"ContainerDied","Data":"9b2d7b84608f1195fd51adfddfacf4d33692276329ff385853e0c3c2a4c4c088"} Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.007582 4906 scope.go:117] "RemoveContainer" containerID="bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.007790 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.017137 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-config-data" (OuterVolumeSpecName: "config-data") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021189 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021214 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vpbf\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-kube-api-access-7vpbf\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021227 4906 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a0222ea-7767-4d08-a6ec-6659a33f9df2-pod-info\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021260 4906 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021274 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9gbw\" (UniqueName: \"kubernetes.io/projected/2e303c56-48b8-4176-99f6-fd0aeb26cf94-kube-api-access-d9gbw\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021285 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021293 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021302 4906 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a0222ea-7767-4d08-a6ec-6659a33f9df2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021311 4906 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.021364 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.026638 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"adbcc53a-3e0f-47b0-a028-01c686a18205","Type":"ContainerDied","Data":"858e0cae8bf896410ba6cada859b5399fbbefca46509879cb8159f22ad27e7f5"} Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.026846 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.027832 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_63999fb5-88e7-4df1-8084-267d0e37ac4c/ovn-northd/0.log" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.027942 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.029776 4906 generic.go:334] "Generic (PLEG): container finished" podID="2e303c56-48b8-4176-99f6-fd0aeb26cf94" containerID="4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb" exitCode=0 Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.029852 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87fd667c9-tzc42" event={"ID":"2e303c56-48b8-4176-99f6-fd0aeb26cf94","Type":"ContainerDied","Data":"4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb"} Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.029886 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87fd667c9-tzc42" event={"ID":"2e303c56-48b8-4176-99f6-fd0aeb26cf94","Type":"ContainerDied","Data":"c8ec7974ff0eb8d271d91f3ef8c4b9b542caa3a6f488bea691468c7520ddd20f"} Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.029937 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87fd667c9-tzc42" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.031350 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data" (OuterVolumeSpecName: "config-data") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.036011 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicana450-account-delete-l2bs5"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.036276 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbicana450-account-delete-l2bs5" podUID="64c1c989-3443-4b80-9786-56b68660655c" containerName="mariadb-account-delete" containerID="cri-o://bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d" gracePeriod=30 Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.039092 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4d677a57-1acf-4627-aa66-f4bee96a2b51","Type":"ContainerDied","Data":"3da95091300888aad706cfc8c6e2300a4e5243f39427c92b2cc8b1420752e895"} Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.039214 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.044215 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a450-account-create-8nn5h"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.047479 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.058506 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-server-conf" (OuterVolumeSpecName: "server-conf") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.058826 4906 scope.go:117] "RemoveContainer" containerID="e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.078097 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.083982 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.092067 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2e303c56-48b8-4176-99f6-fd0aeb26cf94" (UID: "2e303c56-48b8-4176-99f6-fd0aeb26cf94"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.105901 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.120835 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.128843 4906 scope.go:117] "RemoveContainer" containerID="bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.129839 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-combined-ca-bundle\") pod \"63999fb5-88e7-4df1-8084-267d0e37ac4c\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.129882 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-config\") pod \"63999fb5-88e7-4df1-8084-267d0e37ac4c\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.129920 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-northd-tls-certs\") pod \"63999fb5-88e7-4df1-8084-267d0e37ac4c\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.129948 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc5rs\" (UniqueName: \"kubernetes.io/projected/63999fb5-88e7-4df1-8084-267d0e37ac4c-kube-api-access-fc5rs\") pod \"63999fb5-88e7-4df1-8084-267d0e37ac4c\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.129995 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-scripts\") pod \"63999fb5-88e7-4df1-8084-267d0e37ac4c\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130022 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-metrics-certs-tls-certs\") pod \"63999fb5-88e7-4df1-8084-267d0e37ac4c\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130146 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-rundir\") pod \"63999fb5-88e7-4df1-8084-267d0e37ac4c\" (UID: \"63999fb5-88e7-4df1-8084-267d0e37ac4c\") " Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130540 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130558 4906 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-server-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130568 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a0222ea-7767-4d08-a6ec-6659a33f9df2-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130581 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130591 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.130601 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e303c56-48b8-4176-99f6-fd0aeb26cf94-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.131854 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-scripts" (OuterVolumeSpecName: "scripts") pod "63999fb5-88e7-4df1-8084-267d0e37ac4c" (UID: "63999fb5-88e7-4df1-8084-267d0e37ac4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.132333 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af\": container with ID starting with bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af not found: ID does not exist" containerID="bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.132389 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af"} err="failed to get container status \"bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af\": rpc error: code = NotFound desc = could not find container \"bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af\": container with ID starting with bceb4f9fbba2db190344217b97e13a9dd7015da8fa426d0a608b002691fff6af not found: ID does not exist" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.132431 4906 scope.go:117] "RemoveContainer" containerID="e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.132974 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-config" (OuterVolumeSpecName: "config") pod "63999fb5-88e7-4df1-8084-267d0e37ac4c" (UID: "63999fb5-88e7-4df1-8084-267d0e37ac4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.137223 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44\": container with ID starting with e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44 not found: ID does not exist" containerID="e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.137329 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44"} err="failed to get container status \"e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44\": rpc error: code = NotFound desc = could not find container \"e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44\": container with ID starting with e964058df90c9d4edfbbd6adf7ef27d01fbffb9bb1e325b9f5bbf1da7f406c44 not found: ID does not exist" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.137463 4906 scope.go:117] "RemoveContainer" containerID="27497e40b508a21677f03a6794d604b1ad821f2140f70be90b3a8b007d2d719d" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.137226 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "63999fb5-88e7-4df1-8084-267d0e37ac4c" (UID: "63999fb5-88e7-4df1-8084-267d0e37ac4c"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.139084 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63999fb5-88e7-4df1-8084-267d0e37ac4c-kube-api-access-fc5rs" (OuterVolumeSpecName: "kube-api-access-fc5rs") pod "63999fb5-88e7-4df1-8084-267d0e37ac4c" (UID: "63999fb5-88e7-4df1-8084-267d0e37ac4c"). InnerVolumeSpecName "kube-api-access-fc5rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.155698 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.166743 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9a0222ea-7767-4d08-a6ec-6659a33f9df2" (UID: "9a0222ea-7767-4d08-a6ec-6659a33f9df2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.167955 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.191555 4906 scope.go:117] "RemoveContainer" containerID="4e5900f7a006c98c9c587b8cb875127576a79a2f875e56f9d5bcbf31a06d3004" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.207919 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-g6275"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.209852 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63999fb5-88e7-4df1-8084-267d0e37ac4c" (UID: "63999fb5-88e7-4df1-8084-267d0e37ac4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.214758 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-g6275"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.225397 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementb64f-account-delete-576n6"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.227349 4906 scope.go:117] "RemoveContainer" containerID="4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.232347 4906 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.232379 4906 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a0222ea-7767-4d08-a6ec-6659a33f9df2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.232395 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.232405 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.232416 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc5rs\" (UniqueName: \"kubernetes.io/projected/63999fb5-88e7-4df1-8084-267d0e37ac4c-kube-api-access-fc5rs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.232426 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63999fb5-88e7-4df1-8084-267d0e37ac4c-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.237098 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-b64f-account-create-vvmd6"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.243427 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-b64f-account-create-vvmd6"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.248372 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementb64f-account-delete-576n6"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.249187 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "63999fb5-88e7-4df1-8084-267d0e37ac4c" (UID: "63999fb5-88e7-4df1-8084-267d0e37ac4c"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.280591 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "63999fb5-88e7-4df1-8084-267d0e37ac4c" (UID: "63999fb5-88e7-4df1-8084-267d0e37ac4c"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.333616 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-slkpb"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.361467 4906 scope.go:117] "RemoveContainer" containerID="4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb" Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.362092 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.362171 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts podName:01d907ea-d712-40af-b18a-5d55287fbaeb nodeName:}" failed. No retries permitted until 2025-11-23 07:13:55.362141091 +0000 UTC m=+1450.875532394 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts") pod "novacell082ff-account-delete-s8jc5" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb") : configmap "openstack-scripts" not found Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.362285 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb\": container with ID starting with 4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb not found: ID does not exist" containerID="4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.362344 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb"} err="failed to get container status \"4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb\": rpc error: code = NotFound desc = could not find container \"4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb\": container with ID starting with 4708a7e4280a0d2ec8ad946ed0f98947cf242d8294989b64a7353ae766ac59eb not found: ID does not exist" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.362425 4906 scope.go:117] "RemoveContainer" containerID="d83546ce8058eec4db9f5c5b6441b7ce247b7a874eb9b2018665ca353691af54" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.362462 4906 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.362492 4906 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63999fb5-88e7-4df1-8084-267d0e37ac4c-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.363108 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.363278 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts podName:64c1c989-3443-4b80-9786-56b68660655c nodeName:}" failed. No retries permitted until 2025-11-23 07:13:55.3632546 +0000 UTC m=+1450.876645903 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts") pod "barbicana450-account-delete-l2bs5" (UID: "64c1c989-3443-4b80-9786-56b68660655c") : configmap "openstack-scripts" not found Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.458371 4906 scope.go:117] "RemoveContainer" containerID="89f5735b86ee117a6dd64d5fbfd44ffb2a0e87dd340dc5bc635c3b876a04901a" Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.475126 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.475214 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts podName:00991da6-745e-4757-9d89-a20c5b97e38f nodeName:}" failed. No retries permitted until 2025-11-23 07:13:55.475194323 +0000 UTC m=+1450.988585626 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts") pod "neutron9f60-account-delete-vq457" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f") : configmap "openstack-scripts" not found Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.495337 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" path="/var/lib/kubelet/pods/00d152f2-1543-4b14-b214-73747a3f9a26/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.496074 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ab42792-22e3-48ee-8028-3d181679190c" path="/var/lib/kubelet/pods/0ab42792-22e3-48ee-8028-3d181679190c/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.496622 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1642db9a-a07d-475a-8112-518d26323d11" path="/var/lib/kubelet/pods/1642db9a-a07d-475a-8112-518d26323d11/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.498219 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ab3c4dd-5b70-4127-aa8d-570ffcde7477" path="/var/lib/kubelet/pods/1ab3c4dd-5b70-4127-aa8d-570ffcde7477/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.498667 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="212ee22b-c3e7-4d9c-b952-5837b5982c63" path="/var/lib/kubelet/pods/212ee22b-c3e7-4d9c-b952-5837b5982c63/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.499377 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27d994c3-4d19-4084-8c09-d0f2a8d7ede7" path="/var/lib/kubelet/pods/27d994c3-4d19-4084-8c09-d0f2a8d7ede7/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.500406 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c2f5a87-926b-4d48-8e44-ec8dc87f8462" path="/var/lib/kubelet/pods/2c2f5a87-926b-4d48-8e44-ec8dc87f8462/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.501707 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aedece6-346c-4003-8d00-95509a35f23b" path="/var/lib/kubelet/pods/4aedece6-346c-4003-8d00-95509a35f23b/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.502793 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" path="/var/lib/kubelet/pods/4d677a57-1acf-4627-aa66-f4bee96a2b51/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.503781 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6093f84a-c555-4fc4-a207-f54e5ddf2a85" path="/var/lib/kubelet/pods/6093f84a-c555-4fc4-a207-f54e5ddf2a85/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.504264 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" path="/var/lib/kubelet/pods/73149d5c-c16f-4b83-b855-135fc1aab9fe/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.504843 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9fa61a-edd3-476d-8084-10653351773b" path="/var/lib/kubelet/pods/7e9fa61a-edd3-476d-8084-10653351773b/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.505951 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9535028a-afb6-49d6-9bac-541c5324cc55" path="/var/lib/kubelet/pods/9535028a-afb6-49d6-9bac-541c5324cc55/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.506843 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" path="/var/lib/kubelet/pods/a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.508049 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ca0afe-d33d-434b-9182-78ba0588c29e" path="/var/lib/kubelet/pods/a4ca0afe-d33d-434b-9182-78ba0588c29e/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.509498 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adbcc53a-3e0f-47b0-a028-01c686a18205" path="/var/lib/kubelet/pods/adbcc53a-3e0f-47b0-a028-01c686a18205/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.510068 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6673033-6ef9-4a43-90ca-8e1312660b02" path="/var/lib/kubelet/pods/b6673033-6ef9-4a43-90ca-8e1312660b02/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.510555 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da74c07b-2832-4d8e-9508-415fdc3cac71" path="/var/lib/kubelet/pods/da74c07b-2832-4d8e-9508-415fdc3cac71/volumes" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.511997 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-slkpb"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512113 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-46ad-account-create-wc7t2"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512148 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-46ad-account-create-wc7t2"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512161 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder46ad-account-delete-rpn4p"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512175 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder46ad-account-delete-rpn4p"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512187 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-87fd667c9-tzc42"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512221 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-87fd667c9-tzc42"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512234 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512244 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.512256 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-4srzx"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.521844 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-4srzx"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.531373 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-aaf0-account-create-8drwt"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.540733 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanceaaf0-account-delete-j59k7"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.549713 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glanceaaf0-account-delete-j59k7"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.559559 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-aaf0-account-create-8drwt"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.613185 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-4vqvb"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.617159 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-4vqvb"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.638672 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9f60-account-create-ljk4z"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.651494 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron9f60-account-delete-vq457"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.652040 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron9f60-account-delete-vq457" podUID="00991da6-745e-4757-9d89-a20c5b97e38f" containerName="mariadb-account-delete" containerID="cri-o://92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54" gracePeriod=30 Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.655571 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9f60-account-create-ljk4z"] Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.777478 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.778058 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.779871 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.779914 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.780465 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.787026 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.788887 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:51 crc kubenswrapper[4906]: E1123 07:13:51.788922 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.888005 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.943535 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-x42nt"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.964056 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-x42nt"] Nov 23 07:13:51 crc kubenswrapper[4906]: I1123 07:13:51.984610 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-82ff-account-create-5xzpr"] Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.002700 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nqq6\" (UniqueName: \"kubernetes.io/projected/79f707e8-3c2b-4541-94f9-b7cc09cdda72-kube-api-access-7nqq6\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.002864 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-scripts\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.002898 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-combined-ca-bundle\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.002936 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-config-data\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.002994 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-log-httpd\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.003069 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-sg-core-conf-yaml\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.003148 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-run-httpd\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.003190 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-ceilometer-tls-certs\") pod \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\" (UID: \"79f707e8-3c2b-4541-94f9-b7cc09cdda72\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.004168 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.006073 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.007541 4906 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.007567 4906 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79f707e8-3c2b-4541-94f9-b7cc09cdda72-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.009665 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-scripts" (OuterVolumeSpecName: "scripts") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.012910 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f707e8-3c2b-4541-94f9-b7cc09cdda72-kube-api-access-7nqq6" (OuterVolumeSpecName: "kube-api-access-7nqq6") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "kube-api-access-7nqq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.020332 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell082ff-account-delete-s8jc5"] Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.020726 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell082ff-account-delete-s8jc5" podUID="01d907ea-d712-40af-b18a-5d55287fbaeb" containerName="mariadb-account-delete" containerID="cri-o://098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087" gracePeriod=30 Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.027444 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-82ff-account-create-5xzpr"] Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.077742 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.082211 4906 generic.go:334] "Generic (PLEG): container finished" podID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerID="04f1ef97bd7ad66e51909908566ff20e01fcea5f37847fa8e0ca7e76df6a1b60" exitCode=0 Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.082436 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5584fdbb5-9b5k8" event={"ID":"2d3c7e43-fe35-428c-99ae-35245b5d62b6","Type":"ContainerDied","Data":"04f1ef97bd7ad66e51909908566ff20e01fcea5f37847fa8e0ca7e76df6a1b60"} Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.085518 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_63999fb5-88e7-4df1-8084-267d0e37ac4c/ovn-northd/0.log" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.085607 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63999fb5-88e7-4df1-8084-267d0e37ac4c","Type":"ContainerDied","Data":"86dd091db9ec1326251a0c8111f47d74f5b0e80c0c63ef7f27d6f3deb87e779f"} Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.085655 4906 scope.go:117] "RemoveContainer" containerID="631998df3fcc19a80bacb62ef5030d483398f5a45ac22c6b6964329c20351f4a" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.085771 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.085809 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.092856 4906 generic.go:334] "Generic (PLEG): container finished" podID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerID="f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f" exitCode=0 Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.092964 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerDied","Data":"f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f"} Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.093017 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.093016 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79f707e8-3c2b-4541-94f9-b7cc09cdda72","Type":"ContainerDied","Data":"0dbb56eb073b8b5727506abf95dffd56c860f95e0f51c45ee769559f6796acb3"} Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.098836 4906 generic.go:334] "Generic (PLEG): container finished" podID="0ce87b9c-add6-4744-8463-0d268867b9de" containerID="9acde65807d055f979c4c0e765b70fcc9e174d4e93f5057b281b3cd782bbd787" exitCode=0 Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.098908 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" event={"ID":"0ce87b9c-add6-4744-8463-0d268867b9de","Type":"ContainerDied","Data":"9acde65807d055f979c4c0e765b70fcc9e174d4e93f5057b281b3cd782bbd787"} Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.099816 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.101520 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.108773 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-combined-ca-bundle\") pod \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.108873 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data\") pod \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.109003 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3c7e43-fe35-428c-99ae-35245b5d62b6-logs\") pod \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.109118 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf8fg\" (UniqueName: \"kubernetes.io/projected/2d3c7e43-fe35-428c-99ae-35245b5d62b6-kube-api-access-vf8fg\") pod \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.109258 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data-custom\") pod \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\" (UID: \"2d3c7e43-fe35-428c-99ae-35245b5d62b6\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.109944 4906 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.109963 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nqq6\" (UniqueName: \"kubernetes.io/projected/79f707e8-3c2b-4541-94f9-b7cc09cdda72-kube-api-access-7nqq6\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.110039 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.110056 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.110111 4906 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.112011 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d3c7e43-fe35-428c-99ae-35245b5d62b6-logs" (OuterVolumeSpecName: "logs") pod "2d3c7e43-fe35-428c-99ae-35245b5d62b6" (UID: "2d3c7e43-fe35-428c-99ae-35245b5d62b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.114551 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2d3c7e43-fe35-428c-99ae-35245b5d62b6" (UID: "2d3c7e43-fe35-428c-99ae-35245b5d62b6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.131189 4906 scope.go:117] "RemoveContainer" containerID="e6e80c518205ee47f8104f47e81e5d20952e8dfce909932051becb78d5fc0a65" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.132939 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d3c7e43-fe35-428c-99ae-35245b5d62b6-kube-api-access-vf8fg" (OuterVolumeSpecName: "kube-api-access-vf8fg") pod "2d3c7e43-fe35-428c-99ae-35245b5d62b6" (UID: "2d3c7e43-fe35-428c-99ae-35245b5d62b6"). InnerVolumeSpecName "kube-api-access-vf8fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.136707 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.142237 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.157355 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d3c7e43-fe35-428c-99ae-35245b5d62b6" (UID: "2d3c7e43-fe35-428c-99ae-35245b5d62b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.158985 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.168039 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-config-data" (OuterVolumeSpecName: "config-data") pod "79f707e8-3c2b-4541-94f9-b7cc09cdda72" (UID: "79f707e8-3c2b-4541-94f9-b7cc09cdda72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.171469 4906 scope.go:117] "RemoveContainer" containerID="4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.189891 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data" (OuterVolumeSpecName: "config-data") pod "2d3c7e43-fe35-428c-99ae-35245b5d62b6" (UID: "2d3c7e43-fe35-428c-99ae-35245b5d62b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.197108 4906 scope.go:117] "RemoveContainer" containerID="009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.211486 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zld8t\" (UniqueName: \"kubernetes.io/projected/0ce87b9c-add6-4744-8463-0d268867b9de-kube-api-access-zld8t\") pod \"0ce87b9c-add6-4744-8463-0d268867b9de\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.211572 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce87b9c-add6-4744-8463-0d268867b9de-logs\") pod \"0ce87b9c-add6-4744-8463-0d268867b9de\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.211642 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data\") pod \"0ce87b9c-add6-4744-8463-0d268867b9de\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.211752 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data-custom\") pod \"0ce87b9c-add6-4744-8463-0d268867b9de\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.211810 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-combined-ca-bundle\") pod \"0ce87b9c-add6-4744-8463-0d268867b9de\" (UID: \"0ce87b9c-add6-4744-8463-0d268867b9de\") " Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.212137 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79f707e8-3c2b-4541-94f9-b7cc09cdda72-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.212160 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3c7e43-fe35-428c-99ae-35245b5d62b6-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.212170 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf8fg\" (UniqueName: \"kubernetes.io/projected/2d3c7e43-fe35-428c-99ae-35245b5d62b6-kube-api-access-vf8fg\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.212183 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.212192 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.212202 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3c7e43-fe35-428c-99ae-35245b5d62b6-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.213318 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce87b9c-add6-4744-8463-0d268867b9de-logs" (OuterVolumeSpecName: "logs") pod "0ce87b9c-add6-4744-8463-0d268867b9de" (UID: "0ce87b9c-add6-4744-8463-0d268867b9de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.216663 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ce87b9c-add6-4744-8463-0d268867b9de-kube-api-access-zld8t" (OuterVolumeSpecName: "kube-api-access-zld8t") pod "0ce87b9c-add6-4744-8463-0d268867b9de" (UID: "0ce87b9c-add6-4744-8463-0d268867b9de"). InnerVolumeSpecName "kube-api-access-zld8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.217398 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0ce87b9c-add6-4744-8463-0d268867b9de" (UID: "0ce87b9c-add6-4744-8463-0d268867b9de"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.225441 4906 scope.go:117] "RemoveContainer" containerID="f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.235439 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ce87b9c-add6-4744-8463-0d268867b9de" (UID: "0ce87b9c-add6-4744-8463-0d268867b9de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.255081 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data" (OuterVolumeSpecName: "config-data") pod "0ce87b9c-add6-4744-8463-0d268867b9de" (UID: "0ce87b9c-add6-4744-8463-0d268867b9de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.314428 4906 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.314475 4906 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.314487 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ce87b9c-add6-4744-8463-0d268867b9de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.314497 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zld8t\" (UniqueName: \"kubernetes.io/projected/0ce87b9c-add6-4744-8463-0d268867b9de-kube-api-access-zld8t\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.314508 4906 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ce87b9c-add6-4744-8463-0d268867b9de-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.342292 4906 scope.go:117] "RemoveContainer" containerID="f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.364348 4906 scope.go:117] "RemoveContainer" containerID="4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0" Nov 23 07:13:52 crc kubenswrapper[4906]: E1123 07:13:52.364728 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0\": container with ID starting with 4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0 not found: ID does not exist" containerID="4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.364775 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0"} err="failed to get container status \"4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0\": rpc error: code = NotFound desc = could not find container \"4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0\": container with ID starting with 4ef138f5b02f289512d3dacd3d566ecf9f3088d91ea4b4ac403e203eda02ebf0 not found: ID does not exist" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.364807 4906 scope.go:117] "RemoveContainer" containerID="009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5" Nov 23 07:13:52 crc kubenswrapper[4906]: E1123 07:13:52.365192 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5\": container with ID starting with 009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5 not found: ID does not exist" containerID="009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.365229 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5"} err="failed to get container status \"009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5\": rpc error: code = NotFound desc = could not find container \"009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5\": container with ID starting with 009e36da23a8ff26fa0c5a069dbc22828c3c500d412d26a2284c029fb2ffe5e5 not found: ID does not exist" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.365253 4906 scope.go:117] "RemoveContainer" containerID="f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f" Nov 23 07:13:52 crc kubenswrapper[4906]: E1123 07:13:52.365731 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f\": container with ID starting with f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f not found: ID does not exist" containerID="f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.365800 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f"} err="failed to get container status \"f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f\": rpc error: code = NotFound desc = could not find container \"f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f\": container with ID starting with f441924a38cf0e90e83caac835bdb1e695bc5c2cdc96991d276b0ba9c4c4ff2f not found: ID does not exist" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.365846 4906 scope.go:117] "RemoveContainer" containerID="f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df" Nov 23 07:13:52 crc kubenswrapper[4906]: E1123 07:13:52.366245 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df\": container with ID starting with f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df not found: ID does not exist" containerID="f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.366283 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df"} err="failed to get container status \"f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df\": rpc error: code = NotFound desc = could not find container \"f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df\": container with ID starting with f7c1d075891b8e444aa3a944916939e96a446a3dbc062cafa1f4df458f7e13df not found: ID does not exist" Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.468780 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:13:52 crc kubenswrapper[4906]: I1123 07:13:52.481851 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.138986 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" event={"ID":"0ce87b9c-add6-4744-8463-0d268867b9de","Type":"ContainerDied","Data":"a5997aa130cada46fbd35cadfd9b32721747769ca4e08c725a23e4faa8f1983c"} Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.139513 4906 scope.go:117] "RemoveContainer" containerID="9acde65807d055f979c4c0e765b70fcc9e174d4e93f5057b281b3cd782bbd787" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.139329 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f9d47b994-ptdjw" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.152278 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5584fdbb5-9b5k8" event={"ID":"2d3c7e43-fe35-428c-99ae-35245b5d62b6","Type":"ContainerDied","Data":"aa72bd392bf223fd68c4c4e2eb0a7967c0b90a9916d600182f377941466826f0"} Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.152361 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5584fdbb5-9b5k8" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.197016 4906 scope.go:117] "RemoveContainer" containerID="e5685a80b861544fdb6e63f53e8fc51d7652169e4cb7aa78555e20b2db5bcf62" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.197123 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5584fdbb5-9b5k8"] Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.208653 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-5584fdbb5-9b5k8"] Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.214251 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6f9d47b994-ptdjw"] Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.218808 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-6f9d47b994-ptdjw"] Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.242561 4906 scope.go:117] "RemoveContainer" containerID="04f1ef97bd7ad66e51909908566ff20e01fcea5f37847fa8e0ca7e76df6a1b60" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.282612 4906 scope.go:117] "RemoveContainer" containerID="9488749cc23de30cd7d4a1998d3ecd858dad70cda1dd749aa047283ebdd924bf" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.374972 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f3febd-9b3c-4329-a767-16b438b44c91" path="/var/lib/kubelet/pods/02f3febd-9b3c-4329-a767-16b438b44c91/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.375497 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d0ca8b-c938-4cdb-8a50-223295bcc5e4" path="/var/lib/kubelet/pods/08d0ca8b-c938-4cdb-8a50-223295bcc5e4/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.375995 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b28db04-7998-4e4c-80c2-dc2e246f55c5" path="/var/lib/kubelet/pods/0b28db04-7998-4e4c-80c2-dc2e246f55c5/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.376482 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c44516c-f839-4d45-b83d-50a0104d06d9" path="/var/lib/kubelet/pods/0c44516c-f839-4d45-b83d-50a0104d06d9/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.377523 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" path="/var/lib/kubelet/pods/0ce87b9c-add6-4744-8463-0d268867b9de/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.378099 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2510029e-afba-42bb-b00d-8aac204d6b22" path="/var/lib/kubelet/pods/2510029e-afba-42bb-b00d-8aac204d6b22/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.378596 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" path="/var/lib/kubelet/pods/2d3c7e43-fe35-428c-99ae-35245b5d62b6/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.379539 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e0ab9b5-09c4-49b9-a112-76c18f06524d" path="/var/lib/kubelet/pods/2e0ab9b5-09c4-49b9-a112-76c18f06524d/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.380046 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e303c56-48b8-4176-99f6-fd0aeb26cf94" path="/var/lib/kubelet/pods/2e303c56-48b8-4176-99f6-fd0aeb26cf94/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.380490 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4df634a0-630f-46a5-b304-86211157136a" path="/var/lib/kubelet/pods/4df634a0-630f-46a5-b304-86211157136a/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.381527 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" path="/var/lib/kubelet/pods/63999fb5-88e7-4df1-8084-267d0e37ac4c/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.382085 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" path="/var/lib/kubelet/pods/79f707e8-3c2b-4541-94f9-b7cc09cdda72/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.383154 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed64d45-85e5-4a97-beb7-8c31e2a91ca2" path="/var/lib/kubelet/pods/7ed64d45-85e5-4a97-beb7-8c31e2a91ca2/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.383791 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" path="/var/lib/kubelet/pods/9a0222ea-7767-4d08-a6ec-6659a33f9df2/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.384732 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a996f74a-36a3-4c18-a588-ff251e5b40fc" path="/var/lib/kubelet/pods/a996f74a-36a3-4c18-a588-ff251e5b40fc/volumes" Nov 23 07:13:53 crc kubenswrapper[4906]: I1123 07:13:53.385269 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aac4be73-123c-4818-b96d-0a6ea8898fcb" path="/var/lib/kubelet/pods/aac4be73-123c-4818-b96d-0a6ea8898fcb/volumes" Nov 23 07:13:55 crc kubenswrapper[4906]: E1123 07:13:55.368010 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:55 crc kubenswrapper[4906]: E1123 07:13:55.368418 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:55 crc kubenswrapper[4906]: E1123 07:13:55.368438 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts podName:01d907ea-d712-40af-b18a-5d55287fbaeb nodeName:}" failed. No retries permitted until 2025-11-23 07:14:03.368414476 +0000 UTC m=+1458.881805809 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts") pod "novacell082ff-account-delete-s8jc5" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb") : configmap "openstack-scripts" not found Nov 23 07:13:55 crc kubenswrapper[4906]: E1123 07:13:55.368510 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts podName:64c1c989-3443-4b80-9786-56b68660655c nodeName:}" failed. No retries permitted until 2025-11-23 07:14:03.368480847 +0000 UTC m=+1458.881872250 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts") pod "barbicana450-account-delete-l2bs5" (UID: "64c1c989-3443-4b80-9786-56b68660655c") : configmap "openstack-scripts" not found Nov 23 07:13:55 crc kubenswrapper[4906]: E1123 07:13:55.570984 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:13:55 crc kubenswrapper[4906]: E1123 07:13:55.571064 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts podName:00991da6-745e-4757-9d89-a20c5b97e38f nodeName:}" failed. No retries permitted until 2025-11-23 07:14:03.571048065 +0000 UTC m=+1459.084439368 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts") pod "neutron9f60-account-delete-vq457" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f") : configmap "openstack-scripts" not found Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.775609 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.777319 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.777608 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.777642 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.778014 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.779921 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.781403 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:13:56 crc kubenswrapper[4906]: E1123 07:13:56.781443 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.773927 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.774611 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.774916 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.774959 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.775149 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.776832 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.778025 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:01 crc kubenswrapper[4906]: E1123 07:14:01.778061 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:14:03 crc kubenswrapper[4906]: E1123 07:14:03.424248 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:14:03 crc kubenswrapper[4906]: E1123 07:14:03.424663 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts podName:01d907ea-d712-40af-b18a-5d55287fbaeb nodeName:}" failed. No retries permitted until 2025-11-23 07:14:19.424635661 +0000 UTC m=+1474.938026994 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts") pod "novacell082ff-account-delete-s8jc5" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb") : configmap "openstack-scripts" not found Nov 23 07:14:03 crc kubenswrapper[4906]: E1123 07:14:03.424266 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:14:03 crc kubenswrapper[4906]: E1123 07:14:03.424783 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts podName:64c1c989-3443-4b80-9786-56b68660655c nodeName:}" failed. No retries permitted until 2025-11-23 07:14:19.424758314 +0000 UTC m=+1474.938149707 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts") pod "barbicana450-account-delete-l2bs5" (UID: "64c1c989-3443-4b80-9786-56b68660655c") : configmap "openstack-scripts" not found Nov 23 07:14:03 crc kubenswrapper[4906]: E1123 07:14:03.627115 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:14:03 crc kubenswrapper[4906]: E1123 07:14:03.627184 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts podName:00991da6-745e-4757-9d89-a20c5b97e38f nodeName:}" failed. No retries permitted until 2025-11-23 07:14:19.627170608 +0000 UTC m=+1475.140561911 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts") pod "neutron9f60-account-delete-vq457" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f") : configmap "openstack-scripts" not found Nov 23 07:14:05 crc kubenswrapper[4906]: I1123 07:14:05.507619 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7f84fc787f-zskfn" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.155:9696/\": dial tcp 10.217.0.155:9696: connect: connection refused" Nov 23 07:14:05 crc kubenswrapper[4906]: I1123 07:14:05.964053 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.074025 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-combined-ca-bundle\") pod \"e84af371-9517-42aa-943b-a4847f4f8577\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.074474 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-httpd-config\") pod \"e84af371-9517-42aa-943b-a4847f4f8577\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.074548 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-internal-tls-certs\") pod \"e84af371-9517-42aa-943b-a4847f4f8577\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.074595 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-config\") pod \"e84af371-9517-42aa-943b-a4847f4f8577\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.074675 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-public-tls-certs\") pod \"e84af371-9517-42aa-943b-a4847f4f8577\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.074761 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-ovndb-tls-certs\") pod \"e84af371-9517-42aa-943b-a4847f4f8577\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.074787 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plqzn\" (UniqueName: \"kubernetes.io/projected/e84af371-9517-42aa-943b-a4847f4f8577-kube-api-access-plqzn\") pod \"e84af371-9517-42aa-943b-a4847f4f8577\" (UID: \"e84af371-9517-42aa-943b-a4847f4f8577\") " Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.081741 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e84af371-9517-42aa-943b-a4847f4f8577-kube-api-access-plqzn" (OuterVolumeSpecName: "kube-api-access-plqzn") pod "e84af371-9517-42aa-943b-a4847f4f8577" (UID: "e84af371-9517-42aa-943b-a4847f4f8577"). InnerVolumeSpecName "kube-api-access-plqzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.082580 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e84af371-9517-42aa-943b-a4847f4f8577" (UID: "e84af371-9517-42aa-943b-a4847f4f8577"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.116049 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-config" (OuterVolumeSpecName: "config") pod "e84af371-9517-42aa-943b-a4847f4f8577" (UID: "e84af371-9517-42aa-943b-a4847f4f8577"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.117854 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e84af371-9517-42aa-943b-a4847f4f8577" (UID: "e84af371-9517-42aa-943b-a4847f4f8577"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.119007 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e84af371-9517-42aa-943b-a4847f4f8577" (UID: "e84af371-9517-42aa-943b-a4847f4f8577"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.135179 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e84af371-9517-42aa-943b-a4847f4f8577" (UID: "e84af371-9517-42aa-943b-a4847f4f8577"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.153478 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e84af371-9517-42aa-943b-a4847f4f8577" (UID: "e84af371-9517-42aa-943b-a4847f4f8577"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.177256 4906 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.177286 4906 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.177296 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plqzn\" (UniqueName: \"kubernetes.io/projected/e84af371-9517-42aa-943b-a4847f4f8577-kube-api-access-plqzn\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.177309 4906 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.177318 4906 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.177327 4906 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.177336 4906 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e84af371-9517-42aa-943b-a4847f4f8577-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.292653 4906 generic.go:334] "Generic (PLEG): container finished" podID="e84af371-9517-42aa-943b-a4847f4f8577" containerID="dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d" exitCode=0 Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.292778 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f84fc787f-zskfn" event={"ID":"e84af371-9517-42aa-943b-a4847f4f8577","Type":"ContainerDied","Data":"dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d"} Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.292807 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f84fc787f-zskfn" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.292867 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f84fc787f-zskfn" event={"ID":"e84af371-9517-42aa-943b-a4847f4f8577","Type":"ContainerDied","Data":"41bed39932780cf18202c0714f56879367261dc38f9854bf05fdceb5faee7e35"} Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.292911 4906 scope.go:117] "RemoveContainer" containerID="5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.318469 4906 scope.go:117] "RemoveContainer" containerID="dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.340748 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f84fc787f-zskfn"] Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.348474 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7f84fc787f-zskfn"] Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.356943 4906 scope.go:117] "RemoveContainer" containerID="5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390" Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.357323 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390\": container with ID starting with 5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390 not found: ID does not exist" containerID="5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.357376 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390"} err="failed to get container status \"5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390\": rpc error: code = NotFound desc = could not find container \"5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390\": container with ID starting with 5ac555fe8bc778507d1cb56013e14add90df23894943c457745f6f7d106c8390 not found: ID does not exist" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.357408 4906 scope.go:117] "RemoveContainer" containerID="dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d" Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.357634 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d\": container with ID starting with dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d not found: ID does not exist" containerID="dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d" Nov 23 07:14:06 crc kubenswrapper[4906]: I1123 07:14:06.357663 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d"} err="failed to get container status \"dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d\": rpc error: code = NotFound desc = could not find container \"dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d\": container with ID starting with dbdaa71d73f12da46f4094920bc7d8337a53f49629b23f5e423c5d6a5623032d not found: ID does not exist" Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.774476 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.774922 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.775279 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.775304 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.775469 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.779800 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.781474 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:06 crc kubenswrapper[4906]: E1123 07:14:06.781521 4906 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:14:07 crc kubenswrapper[4906]: I1123 07:14:07.367064 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e84af371-9517-42aa-943b-a4847f4f8577" path="/var/lib/kubelet/pods/e84af371-9517-42aa-943b-a4847f4f8577/volumes" Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.775773 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b is running failed: container process not found" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.777707 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b is running failed: container process not found" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.778177 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b is running failed: container process not found" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.778211 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.778518 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.778859 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.779142 4906 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 23 07:14:11 crc kubenswrapper[4906]: E1123 07:14:11.779174 4906 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-ftw6h" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.114589 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ftw6h_e34ddce9-d32e-4c31-9716-86ad39a590e9/ovs-vswitchd/0.log" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.115408 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.204830 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-lib\") pod \"e34ddce9-d32e-4c31-9716-86ad39a590e9\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.204915 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-etc-ovs\") pod \"e34ddce9-d32e-4c31-9716-86ad39a590e9\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.204981 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-log\") pod \"e34ddce9-d32e-4c31-9716-86ad39a590e9\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.205026 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e34ddce9-d32e-4c31-9716-86ad39a590e9-scripts\") pod \"e34ddce9-d32e-4c31-9716-86ad39a590e9\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.205067 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-run\") pod \"e34ddce9-d32e-4c31-9716-86ad39a590e9\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.205101 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r4td\" (UniqueName: \"kubernetes.io/projected/e34ddce9-d32e-4c31-9716-86ad39a590e9-kube-api-access-6r4td\") pod \"e34ddce9-d32e-4c31-9716-86ad39a590e9\" (UID: \"e34ddce9-d32e-4c31-9716-86ad39a590e9\") " Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.206717 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-log" (OuterVolumeSpecName: "var-log") pod "e34ddce9-d32e-4c31-9716-86ad39a590e9" (UID: "e34ddce9-d32e-4c31-9716-86ad39a590e9"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.206767 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-lib" (OuterVolumeSpecName: "var-lib") pod "e34ddce9-d32e-4c31-9716-86ad39a590e9" (UID: "e34ddce9-d32e-4c31-9716-86ad39a590e9"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.206813 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "e34ddce9-d32e-4c31-9716-86ad39a590e9" (UID: "e34ddce9-d32e-4c31-9716-86ad39a590e9"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.206901 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-run" (OuterVolumeSpecName: "var-run") pod "e34ddce9-d32e-4c31-9716-86ad39a590e9" (UID: "e34ddce9-d32e-4c31-9716-86ad39a590e9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.207558 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e34ddce9-d32e-4c31-9716-86ad39a590e9-scripts" (OuterVolumeSpecName: "scripts") pod "e34ddce9-d32e-4c31-9716-86ad39a590e9" (UID: "e34ddce9-d32e-4c31-9716-86ad39a590e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.211075 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e34ddce9-d32e-4c31-9716-86ad39a590e9-kube-api-access-6r4td" (OuterVolumeSpecName: "kube-api-access-6r4td") pod "e34ddce9-d32e-4c31-9716-86ad39a590e9" (UID: "e34ddce9-d32e-4c31-9716-86ad39a590e9"). InnerVolumeSpecName "kube-api-access-6r4td". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.306977 4906 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-lib\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.307333 4906 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-etc-ovs\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.307343 4906 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-log\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.307353 4906 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e34ddce9-d32e-4c31-9716-86ad39a590e9-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.307365 4906 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e34ddce9-d32e-4c31-9716-86ad39a590e9-var-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.307380 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r4td\" (UniqueName: \"kubernetes.io/projected/e34ddce9-d32e-4c31-9716-86ad39a590e9-kube-api-access-6r4td\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.362459 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ftw6h_e34ddce9-d32e-4c31-9716-86ad39a590e9/ovs-vswitchd/0.log" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.365281 4906 generic.go:334] "Generic (PLEG): container finished" podID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" exitCode=137 Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.365338 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ftw6h" event={"ID":"e34ddce9-d32e-4c31-9716-86ad39a590e9","Type":"ContainerDied","Data":"45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b"} Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.365375 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ftw6h" event={"ID":"e34ddce9-d32e-4c31-9716-86ad39a590e9","Type":"ContainerDied","Data":"905ba1126072f9b90864ddc6b6599d63d7f49861d6bb3ab0ed2379c179665006"} Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.365398 4906 scope.go:117] "RemoveContainer" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.365597 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ftw6h" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.408572 4906 scope.go:117] "RemoveContainer" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.411850 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-ftw6h"] Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.417854 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-ftw6h"] Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.435536 4906 scope.go:117] "RemoveContainer" containerID="85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.468339 4906 scope.go:117] "RemoveContainer" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" Nov 23 07:14:12 crc kubenswrapper[4906]: E1123 07:14:12.469284 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b\": container with ID starting with 45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b not found: ID does not exist" containerID="45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.469338 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b"} err="failed to get container status \"45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b\": rpc error: code = NotFound desc = could not find container \"45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b\": container with ID starting with 45694ebf0f6010addab66e1cd4caa43ca2f239c9bc496ccd478ff2030eaea35b not found: ID does not exist" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.469371 4906 scope.go:117] "RemoveContainer" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" Nov 23 07:14:12 crc kubenswrapper[4906]: E1123 07:14:12.470112 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4\": container with ID starting with cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 not found: ID does not exist" containerID="cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.470203 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4"} err="failed to get container status \"cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4\": rpc error: code = NotFound desc = could not find container \"cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4\": container with ID starting with cd12aa446af574aec335067a2b9d7bf753b964b3bb1bf219c8aa58a8a5b4bcd4 not found: ID does not exist" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.470227 4906 scope.go:117] "RemoveContainer" containerID="85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0" Nov 23 07:14:12 crc kubenswrapper[4906]: E1123 07:14:12.470828 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0\": container with ID starting with 85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0 not found: ID does not exist" containerID="85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0" Nov 23 07:14:12 crc kubenswrapper[4906]: I1123 07:14:12.470934 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0"} err="failed to get container status \"85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0\": rpc error: code = NotFound desc = could not find container \"85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0\": container with ID starting with 85b0a6415da6b5488af2df6e6e6ff64b6c4e8ad7a960ff46513ddb114509c3e0 not found: ID does not exist" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.382997 4906 generic.go:334] "Generic (PLEG): container finished" podID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerID="0f3631e076f67c96972eae5e4520d40c2b4e9bafcc51ca7093e8a282b2d1ab2d" exitCode=137 Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.384736 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" path="/var/lib/kubelet/pods/e34ddce9-d32e-4c31-9716-86ad39a590e9/volumes" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.386351 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"0f3631e076f67c96972eae5e4520d40c2b4e9bafcc51ca7093e8a282b2d1ab2d"} Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.559527 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.726010 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-cache\") pod \"c4ec7c62-702e-4794-81fc-64c72f72cb12\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.726116 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"c4ec7c62-702e-4794-81fc-64c72f72cb12\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.726151 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-lock\") pod \"c4ec7c62-702e-4794-81fc-64c72f72cb12\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.726344 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") pod \"c4ec7c62-702e-4794-81fc-64c72f72cb12\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.726375 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lsl4\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-kube-api-access-7lsl4\") pod \"c4ec7c62-702e-4794-81fc-64c72f72cb12\" (UID: \"c4ec7c62-702e-4794-81fc-64c72f72cb12\") " Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.727251 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-lock" (OuterVolumeSpecName: "lock") pod "c4ec7c62-702e-4794-81fc-64c72f72cb12" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.727389 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-cache" (OuterVolumeSpecName: "cache") pod "c4ec7c62-702e-4794-81fc-64c72f72cb12" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.732785 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "swift") pod "c4ec7c62-702e-4794-81fc-64c72f72cb12" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.734908 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-kube-api-access-7lsl4" (OuterVolumeSpecName: "kube-api-access-7lsl4") pod "c4ec7c62-702e-4794-81fc-64c72f72cb12" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12"). InnerVolumeSpecName "kube-api-access-7lsl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.739907 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c4ec7c62-702e-4794-81fc-64c72f72cb12" (UID: "c4ec7c62-702e-4794-81fc-64c72f72cb12"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.828340 4906 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-cache\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.828401 4906 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.828413 4906 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4ec7c62-702e-4794-81fc-64c72f72cb12-lock\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.828423 4906 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.828434 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lsl4\" (UniqueName: \"kubernetes.io/projected/c4ec7c62-702e-4794-81fc-64c72f72cb12-kube-api-access-7lsl4\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.844317 4906 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 23 07:14:13 crc kubenswrapper[4906]: I1123 07:14:13.930015 4906 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.404238 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4ec7c62-702e-4794-81fc-64c72f72cb12","Type":"ContainerDied","Data":"90b0d500e80584b17478edcd992a6916fcff5fcf37c96acd17e254ed07ee3bd0"} Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.404698 4906 scope.go:117] "RemoveContainer" containerID="0f3631e076f67c96972eae5e4520d40c2b4e9bafcc51ca7093e8a282b2d1ab2d" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.404588 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.432938 4906 scope.go:117] "RemoveContainer" containerID="c1437bfa849a403a91c6260f5cd5aec30c660352ad87626d59e56d5664d49e22" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.455900 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.466279 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.468138 4906 scope.go:117] "RemoveContainer" containerID="18a0cbb3bc39086918add992d8b7e36417c2a512ce34ad58af2473769333d276" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.504818 4906 scope.go:117] "RemoveContainer" containerID="7310698561fe415b603d4b23e1ca0b97cb10e0a5ab1b9e756910372da0037e1f" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.528014 4906 scope.go:117] "RemoveContainer" containerID="52c1da52f5e6b3969985d8979a7d057a2abc04e85e14b228a48cb576460bae01" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.547527 4906 scope.go:117] "RemoveContainer" containerID="28356b85be5f4f2918de3e09fcd6083deb2b49f2e911a40c85312e0eebf206a8" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.572133 4906 scope.go:117] "RemoveContainer" containerID="14e047c6504017d7ca6ee4d097e4dc0c1a3d8b79305ee2e36580fe0c8238c99b" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.603223 4906 scope.go:117] "RemoveContainer" containerID="627197269c9d973c5e754357b5bea2cd6202f4ed8b342e5be634b4ffe2fc24f4" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.635858 4906 scope.go:117] "RemoveContainer" containerID="59f81af64ef288c720feec11a22a60062d378a09a432fe305317b0c5b6c0c0a3" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.660455 4906 scope.go:117] "RemoveContainer" containerID="d7765fe2025af755e1f8644c577f06f2002c8f768fc520d7b9ae7dfe318bd069" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.684665 4906 scope.go:117] "RemoveContainer" containerID="0c11306e0c7841c07b12efafe09a24ad91a89cee793dba18a9f2876f0d47db2f" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.702815 4906 scope.go:117] "RemoveContainer" containerID="11af2df38e7407624c492ea4718898da13ce0775329468c7c775d8b83ef7a458" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.722349 4906 scope.go:117] "RemoveContainer" containerID="29fc0d262044ca645fb4133e01cbabc9655fe5b59a91b3028e0d08bcd5278256" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.740236 4906 scope.go:117] "RemoveContainer" containerID="8bed0393fc8307e59d410b24ec754b21fe7d3bd1f3e83cd8d6f028d1a8f06cdd" Nov 23 07:14:14 crc kubenswrapper[4906]: I1123 07:14:14.762372 4906 scope.go:117] "RemoveContainer" containerID="4f7149c3b0589b8222d668111ff70a5b576f33041ccbfe291e9f62085cc511d3" Nov 23 07:14:15 crc kubenswrapper[4906]: I1123 07:14:15.370785 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" path="/var/lib/kubelet/pods/c4ec7c62-702e-4794-81fc-64c72f72cb12/volumes" Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.012197 4906 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poda7788825-483c-46f2-a975-239254435019"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poda7788825-483c-46f2-a975-239254435019] : Timed out while waiting for systemd to remove kubepods-besteffort-poda7788825_483c_46f2_a975_239254435019.slice" Nov 23 07:14:18 crc kubenswrapper[4906]: E1123 07:14:18.012762 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort poda7788825-483c-46f2-a975-239254435019] : unable to destroy cgroup paths for cgroup [kubepods besteffort poda7788825-483c-46f2-a975-239254435019] : Timed out while waiting for systemd to remove kubepods-besteffort-poda7788825_483c_46f2_a975_239254435019.slice" pod="openstack/cinder-scheduler-0" podUID="a7788825-483c-46f2-a975-239254435019" Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.018789 4906 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod51dd433b-f1a2-422a-bf5b-643e150c3bca"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod51dd433b-f1a2-422a-bf5b-643e150c3bca] : Timed out while waiting for systemd to remove kubepods-besteffort-pod51dd433b_f1a2_422a_bf5b_643e150c3bca.slice" Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.020971 4906 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod0a95e1df-3578-4430-b0ac-befb54723517"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod0a95e1df-3578-4430-b0ac-befb54723517] : Timed out while waiting for systemd to remove kubepods-besteffort-pod0a95e1df_3578_4430_b0ac_befb54723517.slice" Nov 23 07:14:18 crc kubenswrapper[4906]: E1123 07:14:18.021031 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod0a95e1df-3578-4430-b0ac-befb54723517] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod0a95e1df-3578-4430-b0ac-befb54723517] : Timed out while waiting for systemd to remove kubepods-besteffort-pod0a95e1df_3578_4430_b0ac_befb54723517.slice" pod="openstack/openstack-cell1-galera-0" podUID="0a95e1df-3578-4430-b0ac-befb54723517" Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.449708 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.449713 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.520350 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.531248 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.537319 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:14:18 crc kubenswrapper[4906]: I1123 07:14:18.542989 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:14:19 crc kubenswrapper[4906]: I1123 07:14:19.372306 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a95e1df-3578-4430-b0ac-befb54723517" path="/var/lib/kubelet/pods/0a95e1df-3578-4430-b0ac-befb54723517/volumes" Nov 23 07:14:19 crc kubenswrapper[4906]: I1123 07:14:19.373858 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7788825-483c-46f2-a975-239254435019" path="/var/lib/kubelet/pods/a7788825-483c-46f2-a975-239254435019/volumes" Nov 23 07:14:19 crc kubenswrapper[4906]: E1123 07:14:19.516965 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:14:19 crc kubenswrapper[4906]: E1123 07:14:19.517084 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:14:19 crc kubenswrapper[4906]: E1123 07:14:19.517135 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts podName:01d907ea-d712-40af-b18a-5d55287fbaeb nodeName:}" failed. No retries permitted until 2025-11-23 07:14:51.517100417 +0000 UTC m=+1507.030491760 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts") pod "novacell082ff-account-delete-s8jc5" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb") : configmap "openstack-scripts" not found Nov 23 07:14:19 crc kubenswrapper[4906]: E1123 07:14:19.517176 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts podName:64c1c989-3443-4b80-9786-56b68660655c nodeName:}" failed. No retries permitted until 2025-11-23 07:14:51.517160438 +0000 UTC m=+1507.030551871 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts") pod "barbicana450-account-delete-l2bs5" (UID: "64c1c989-3443-4b80-9786-56b68660655c") : configmap "openstack-scripts" not found Nov 23 07:14:19 crc kubenswrapper[4906]: E1123 07:14:19.720714 4906 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 23 07:14:19 crc kubenswrapper[4906]: E1123 07:14:19.720822 4906 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts podName:00991da6-745e-4757-9d89-a20c5b97e38f nodeName:}" failed. No retries permitted until 2025-11-23 07:14:51.720802194 +0000 UTC m=+1507.234193497 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts") pod "neutron9f60-account-delete-vq457" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f") : configmap "openstack-scripts" not found Nov 23 07:14:20 crc kubenswrapper[4906]: I1123 07:14:20.945497 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:14:20 crc kubenswrapper[4906]: I1123 07:14:20.945580 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:14:20 crc kubenswrapper[4906]: I1123 07:14:20.945641 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:14:20 crc kubenswrapper[4906]: I1123 07:14:20.946213 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"efb82b69034455635a0b7ce63f8c53d1bea8631681ec6e17e626273b28cdbdd5"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:14:20 crc kubenswrapper[4906]: I1123 07:14:20.946267 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://efb82b69034455635a0b7ce63f8c53d1bea8631681ec6e17e626273b28cdbdd5" gracePeriod=600 Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.414212 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.488917 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="efb82b69034455635a0b7ce63f8c53d1bea8631681ec6e17e626273b28cdbdd5" exitCode=0 Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.489029 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"efb82b69034455635a0b7ce63f8c53d1bea8631681ec6e17e626273b28cdbdd5"} Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.489117 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7"} Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.489149 4906 scope.go:117] "RemoveContainer" containerID="cde0743cc2b7d9547c3784b89c951ad5af5f1a5ec9c54292ba0b98af998c4c47" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.491574 4906 generic.go:334] "Generic (PLEG): container finished" podID="64c1c989-3443-4b80-9786-56b68660655c" containerID="bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d" exitCode=137 Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.491762 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicana450-account-delete-l2bs5" event={"ID":"64c1c989-3443-4b80-9786-56b68660655c","Type":"ContainerDied","Data":"bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d"} Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.492252 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicana450-account-delete-l2bs5" event={"ID":"64c1c989-3443-4b80-9786-56b68660655c","Type":"ContainerDied","Data":"65f008398b0a61b9837f9c35e83f9327847033e6ccd77632fde49a148a93fb1c"} Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.492478 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicana450-account-delete-l2bs5" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.525815 4906 scope.go:117] "RemoveContainer" containerID="bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.549896 4906 scope.go:117] "RemoveContainer" containerID="bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d" Nov 23 07:14:21 crc kubenswrapper[4906]: E1123 07:14:21.550663 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d\": container with ID starting with bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d not found: ID does not exist" containerID="bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.550733 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d"} err="failed to get container status \"bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d\": rpc error: code = NotFound desc = could not find container \"bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d\": container with ID starting with bb65b9c85b5260d0fd5b4c4472f5aae26a6a3ed074b50b7e2efb7b1b874fad7d not found: ID does not exist" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.555239 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts\") pod \"64c1c989-3443-4b80-9786-56b68660655c\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.555430 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsw9k\" (UniqueName: \"kubernetes.io/projected/64c1c989-3443-4b80-9786-56b68660655c-kube-api-access-rsw9k\") pod \"64c1c989-3443-4b80-9786-56b68660655c\" (UID: \"64c1c989-3443-4b80-9786-56b68660655c\") " Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.556638 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64c1c989-3443-4b80-9786-56b68660655c" (UID: "64c1c989-3443-4b80-9786-56b68660655c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.563501 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c1c989-3443-4b80-9786-56b68660655c-kube-api-access-rsw9k" (OuterVolumeSpecName: "kube-api-access-rsw9k") pod "64c1c989-3443-4b80-9786-56b68660655c" (UID: "64c1c989-3443-4b80-9786-56b68660655c"). InnerVolumeSpecName "kube-api-access-rsw9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.658160 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsw9k\" (UniqueName: \"kubernetes.io/projected/64c1c989-3443-4b80-9786-56b68660655c-kube-api-access-rsw9k\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.658212 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64c1c989-3443-4b80-9786-56b68660655c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.834921 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicana450-account-delete-l2bs5"] Nov 23 07:14:21 crc kubenswrapper[4906]: I1123 07:14:21.844728 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicana450-account-delete-l2bs5"] Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.027359 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.172669 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqcfw\" (UniqueName: \"kubernetes.io/projected/00991da6-745e-4757-9d89-a20c5b97e38f-kube-api-access-rqcfw\") pod \"00991da6-745e-4757-9d89-a20c5b97e38f\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.172774 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts\") pod \"00991da6-745e-4757-9d89-a20c5b97e38f\" (UID: \"00991da6-745e-4757-9d89-a20c5b97e38f\") " Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.173835 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00991da6-745e-4757-9d89-a20c5b97e38f" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.181569 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00991da6-745e-4757-9d89-a20c5b97e38f-kube-api-access-rqcfw" (OuterVolumeSpecName: "kube-api-access-rqcfw") pod "00991da6-745e-4757-9d89-a20c5b97e38f" (UID: "00991da6-745e-4757-9d89-a20c5b97e38f"). InnerVolumeSpecName "kube-api-access-rqcfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.275192 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqcfw\" (UniqueName: \"kubernetes.io/projected/00991da6-745e-4757-9d89-a20c5b97e38f-kube-api-access-rqcfw\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.275228 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00991da6-745e-4757-9d89-a20c5b97e38f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.422342 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.482772 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts\") pod \"01d907ea-d712-40af-b18a-5d55287fbaeb\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.482918 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxrdn\" (UniqueName: \"kubernetes.io/projected/01d907ea-d712-40af-b18a-5d55287fbaeb-kube-api-access-gxrdn\") pod \"01d907ea-d712-40af-b18a-5d55287fbaeb\" (UID: \"01d907ea-d712-40af-b18a-5d55287fbaeb\") " Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.484053 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01d907ea-d712-40af-b18a-5d55287fbaeb" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.491269 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01d907ea-d712-40af-b18a-5d55287fbaeb-kube-api-access-gxrdn" (OuterVolumeSpecName: "kube-api-access-gxrdn") pod "01d907ea-d712-40af-b18a-5d55287fbaeb" (UID: "01d907ea-d712-40af-b18a-5d55287fbaeb"). InnerVolumeSpecName "kube-api-access-gxrdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.524736 4906 generic.go:334] "Generic (PLEG): container finished" podID="01d907ea-d712-40af-b18a-5d55287fbaeb" containerID="098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087" exitCode=137 Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.524831 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell082ff-account-delete-s8jc5" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.524922 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell082ff-account-delete-s8jc5" event={"ID":"01d907ea-d712-40af-b18a-5d55287fbaeb","Type":"ContainerDied","Data":"098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087"} Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.525012 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell082ff-account-delete-s8jc5" event={"ID":"01d907ea-d712-40af-b18a-5d55287fbaeb","Type":"ContainerDied","Data":"95317f08311eec38eda1ec3bca3eb11cc85b1f3aeb23b3c83a54d447ed5285e0"} Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.525045 4906 scope.go:117] "RemoveContainer" containerID="098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.527365 4906 generic.go:334] "Generic (PLEG): container finished" podID="00991da6-745e-4757-9d89-a20c5b97e38f" containerID="92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54" exitCode=137 Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.527434 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron9f60-account-delete-vq457" event={"ID":"00991da6-745e-4757-9d89-a20c5b97e38f","Type":"ContainerDied","Data":"92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54"} Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.527529 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron9f60-account-delete-vq457" event={"ID":"00991da6-745e-4757-9d89-a20c5b97e38f","Type":"ContainerDied","Data":"ed471134383b6a342d71e4523be2b2493de4106ca7705b9219d7f8842e8521cc"} Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.527503 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron9f60-account-delete-vq457" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.562598 4906 scope.go:117] "RemoveContainer" containerID="098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087" Nov 23 07:14:22 crc kubenswrapper[4906]: E1123 07:14:22.563338 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087\": container with ID starting with 098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087 not found: ID does not exist" containerID="098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.563375 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087"} err="failed to get container status \"098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087\": rpc error: code = NotFound desc = could not find container \"098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087\": container with ID starting with 098180685ad68a852cd1afa8661ed5f71115b65b9253747754cd6290a7b5d087 not found: ID does not exist" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.563405 4906 scope.go:117] "RemoveContainer" containerID="92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.577906 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell082ff-account-delete-s8jc5"] Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.582929 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell082ff-account-delete-s8jc5"] Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.586231 4906 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01d907ea-d712-40af-b18a-5d55287fbaeb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.586291 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxrdn\" (UniqueName: \"kubernetes.io/projected/01d907ea-d712-40af-b18a-5d55287fbaeb-kube-api-access-gxrdn\") on node \"crc\" DevicePath \"\"" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.586931 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron9f60-account-delete-vq457"] Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.594886 4906 scope.go:117] "RemoveContainer" containerID="92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.595349 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron9f60-account-delete-vq457"] Nov 23 07:14:22 crc kubenswrapper[4906]: E1123 07:14:22.595597 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54\": container with ID starting with 92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54 not found: ID does not exist" containerID="92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54" Nov 23 07:14:22 crc kubenswrapper[4906]: I1123 07:14:22.595657 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54"} err="failed to get container status \"92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54\": rpc error: code = NotFound desc = could not find container \"92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54\": container with ID starting with 92dd3cf507bd884f720f207375729161ca41aa5044976bc20dfaec01113d7a54 not found: ID does not exist" Nov 23 07:14:23 crc kubenswrapper[4906]: I1123 07:14:23.365848 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00991da6-745e-4757-9d89-a20c5b97e38f" path="/var/lib/kubelet/pods/00991da6-745e-4757-9d89-a20c5b97e38f/volumes" Nov 23 07:14:23 crc kubenswrapper[4906]: I1123 07:14:23.367016 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01d907ea-d712-40af-b18a-5d55287fbaeb" path="/var/lib/kubelet/pods/01d907ea-d712-40af-b18a-5d55287fbaeb/volumes" Nov 23 07:14:23 crc kubenswrapper[4906]: I1123 07:14:23.367790 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c1c989-3443-4b80-9786-56b68660655c" path="/var/lib/kubelet/pods/64c1c989-3443-4b80-9786-56b68660655c/volumes" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.690219 4906 scope.go:117] "RemoveContainer" containerID="adb454b88595fdda4ef098a37bd75a106172393c0263a8fabed2363a86dda08c" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.728450 4906 scope.go:117] "RemoveContainer" containerID="717d5d2bacd9b96545d2a48c905f3b13c01a02ade9d241f21d8207a70788d9a9" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.766329 4906 scope.go:117] "RemoveContainer" containerID="9e4e6012c2cef7cf10c2d0a9aa750a115c7cac5349940c88ef4bec142dc282da" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.802313 4906 scope.go:117] "RemoveContainer" containerID="bf5e01bb5e89dd0c74104dbe793e0f8882ba08a29b84959fea106fea381a3521" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.837272 4906 scope.go:117] "RemoveContainer" containerID="8f173df3c1376722f31ab2e49d813520a0aa789fb2f6fe19a15bde7d99bd2b3b" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.859592 4906 scope.go:117] "RemoveContainer" containerID="4defce0e3fb0237aa4ffc1b4b5e58328d1a52cca3c133f1db7f360d81d75d7ac" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.906749 4906 scope.go:117] "RemoveContainer" containerID="3acf2c2a799d15e29b4ea27d19732e6c4972f398d8d2402a22dee98452de758a" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.928396 4906 scope.go:117] "RemoveContainer" containerID="3d34a0cfbf6cce02048082a86c433bac46bdbc467dfe235995d608ebba0146d0" Nov 23 07:14:53 crc kubenswrapper[4906]: I1123 07:14:53.949282 4906 scope.go:117] "RemoveContainer" containerID="89459c3df93e572c8aa54d6cdf68fb56123719eb60818d2b1a1c63833b8221fa" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.154653 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj"] Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156356 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156382 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156421 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156431 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-server" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156448 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab42792-22e3-48ee-8028-3d181679190c" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156458 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab42792-22e3-48ee-8028-3d181679190c" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156479 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerName="galera" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156489 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerName="galera" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156505 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7788825-483c-46f2-a975-239254435019" containerName="probe" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156514 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7788825-483c-46f2-a975-239254435019" containerName="probe" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156530 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e303c56-48b8-4176-99f6-fd0aeb26cf94" containerName="keystone-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156540 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e303c56-48b8-4176-99f6-fd0aeb26cf94" containerName="keystone-api" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156553 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerName="setup-container" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156563 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerName="setup-container" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156578 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6093f84a-c555-4fc4-a207-f54e5ddf2a85" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156587 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="6093f84a-c555-4fc4-a207-f54e5ddf2a85" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156604 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00991da6-745e-4757-9d89-a20c5b97e38f" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156615 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="00991da6-745e-4757-9d89-a20c5b97e38f" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156629 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156639 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156648 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156657 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156672 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="sg-core" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156705 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="sg-core" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156714 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156723 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156740 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156750 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156764 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cee328a-b746-486d-850b-148ca1e66eea" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156773 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cee328a-b746-486d-850b-148ca1e66eea" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156784 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156795 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156810 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da74c07b-2832-4d8e-9508-415fdc3cac71" containerName="nova-scheduler-scheduler" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156819 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="da74c07b-2832-4d8e-9508-415fdc3cac71" containerName="nova-scheduler-scheduler" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156832 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="212ee22b-c3e7-4d9c-b952-5837b5982c63" containerName="nova-cell1-conductor-conductor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156842 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="212ee22b-c3e7-4d9c-b952-5837b5982c63" containerName="nova-cell1-conductor-conductor" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156858 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acf9ff7-0012-4e1a-b940-037bebcc4c88" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156866 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acf9ff7-0012-4e1a-b940-037bebcc4c88" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156875 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156883 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156893 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-expirer" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156905 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-expirer" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156918 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server-init" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156929 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server-init" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156943 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156950 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156967 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.156975 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.156992 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157005 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157019 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157027 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157041 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerName="rabbitmq" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157049 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerName="rabbitmq" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157060 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157069 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-api" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157082 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df634a0-630f-46a5-b304-86211157136a" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157091 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df634a0-630f-46a5-b304-86211157136a" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157103 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c1c989-3443-4b80-9786-56b68660655c" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157111 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c1c989-3443-4b80-9786-56b68660655c" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157122 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157130 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-api" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157139 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-reaper" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157147 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-reaper" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157158 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157166 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157182 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="proxy-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157193 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="proxy-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157209 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="ovn-northd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157218 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="ovn-northd" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157232 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157240 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-api" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157249 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157258 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157272 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157281 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157293 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a95e1df-3578-4430-b0ac-befb54723517" containerName="galera" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157301 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a95e1df-3578-4430-b0ac-befb54723517" containerName="galera" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157313 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-updater" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157321 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-updater" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157331 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157342 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157355 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157363 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157374 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d994c3-4d19-4084-8c09-d0f2a8d7ede7" containerName="memcached" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157383 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d994c3-4d19-4084-8c09-d0f2a8d7ede7" containerName="memcached" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157393 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01d907ea-d712-40af-b18a-5d55287fbaeb" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157402 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="01d907ea-d712-40af-b18a-5d55287fbaeb" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157414 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerName="rabbitmq" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157422 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerName="rabbitmq" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157436 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-updater" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157446 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-updater" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157460 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerName="mysql-bootstrap" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157469 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerName="mysql-bootstrap" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157483 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerName="setup-container" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157492 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerName="setup-container" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157500 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157508 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-server" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157520 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="rsync" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157530 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="rsync" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157547 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157556 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157564 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157573 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-server" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157585 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157593 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157607 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="ovsdbserver-sb" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157615 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="ovsdbserver-sb" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157626 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157636 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157646 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157656 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157671 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157701 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157714 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-central-agent" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157725 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-central-agent" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157740 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157750 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157764 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157775 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157792 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerName="init" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157801 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerName="init" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157817 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157827 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157838 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157848 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157863 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-metadata" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157875 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-metadata" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157887 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0ab9b5-09c4-49b9-a112-76c18f06524d" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157897 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0ab9b5-09c4-49b9-a112-76c18f06524d" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157908 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157918 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-server" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157933 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157942 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157955 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157965 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.157976 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.157984 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158001 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="ovsdbserver-nb" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158011 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="ovsdbserver-nb" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158025 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158035 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener-log" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158046 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a95e1df-3578-4430-b0ac-befb54723517" containerName="mysql-bootstrap" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158056 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a95e1df-3578-4430-b0ac-befb54723517" containerName="mysql-bootstrap" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158067 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="swift-recon-cron" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158077 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="swift-recon-cron" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158087 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab3c4dd-5b70-4127-aa8d-570ffcde7477" containerName="nova-cell0-conductor-conductor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158097 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab3c4dd-5b70-4127-aa8d-570ffcde7477" containerName="nova-cell0-conductor-conductor" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158110 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-notification-agent" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158120 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-notification-agent" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158131 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerName="dnsmasq-dns" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158144 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerName="dnsmasq-dns" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158156 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7788825-483c-46f2-a975-239254435019" containerName="cinder-scheduler" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158165 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7788825-483c-46f2-a975-239254435019" containerName="cinder-scheduler" Nov 23 07:15:00 crc kubenswrapper[4906]: E1123 07:15:00.158176 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9535028a-afb6-49d6-9bac-541c5324cc55" containerName="kube-state-metrics" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158185 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9535028a-afb6-49d6-9bac-541c5324cc55" containerName="kube-state-metrics" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158415 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158441 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158460 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158475 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-updater" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158490 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158503 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="swift-recon-cron" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158514 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158523 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="rsync" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158538 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="sg-core" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158556 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1642db9a-a07d-475a-8112-518d26323d11" containerName="glance-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158569 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="ovsdbserver-sb" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158583 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158592 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="00991da6-745e-4757-9d89-a20c5b97e38f" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158603 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-updater" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158612 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d994c3-4d19-4084-8c09-d0f2a8d7ede7" containerName="memcached" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158621 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158634 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7788825-483c-46f2-a975-239254435019" containerName="cinder-scheduler" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158642 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="proxy-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158651 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158665 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4121204-6477-44c6-b669-fc7497fa908d" containerName="proxy-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158675 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cee328a-b746-486d-850b-148ca1e66eea" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158704 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158718 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158730 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158740 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158754 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="51dd433b-f1a2-422a-bf5b-643e150c3bca" containerName="ovsdbserver-nb" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158763 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="212ee22b-c3e7-4d9c-b952-5837b5982c63" containerName="nova-cell1-conductor-conductor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158772 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4df634a0-630f-46a5-b304-86211157136a" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158781 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e0ab9b5-09c4-49b9-a112-76c18f06524d" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158790 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d152f2-1543-4b14-b214-73747a3f9a26" containerName="nova-metadata-metadata" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158805 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158814 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c2214f-2ce4-4c2f-93ed-4a20cdb91a4d" containerName="glance-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158828 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="da74c07b-2832-4d8e-9508-415fdc3cac71" containerName="nova-scheduler-scheduler" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158836 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7788825-483c-46f2-a975-239254435019" containerName="probe" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158848 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-expirer" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158859 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d3c7e43-fe35-428c-99ae-35245b5d62b6" containerName="barbican-worker-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158871 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0222ea-7767-4d08-a6ec-6659a33f9df2" containerName="rabbitmq" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158883 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e303c56-48b8-4176-99f6-fd0aeb26cf94" containerName="keystone-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158891 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="6093f84a-c555-4fc4-a207-f54e5ddf2a85" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158903 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="3acf9ff7-0012-4e1a-b940-037bebcc4c88" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158915 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="ovn-northd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158926 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158936 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="adbcc53a-3e0f-47b0-a028-01c686a18205" containerName="galera" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158946 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="9535028a-afb6-49d6-9bac-541c5324cc55" containerName="kube-state-metrics" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158960 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158974 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.158986 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ab42792-22e3-48ee-8028-3d181679190c" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161636 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovs-vswitchd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161662 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a95e1df-3578-4430-b0ac-befb54723517" containerName="galera" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161696 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-replicator" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161709 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-notification-agent" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161730 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40394f2-6a98-49df-813a-17db4ced9cd2" containerName="dnsmasq-dns" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161742 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f707e8-3c2b-4541-94f9-b7cc09cdda72" containerName="ceilometer-central-agent" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161753 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161766 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34ddce9-d32e-4c31-9716-86ad39a590e9" containerName="ovsdb-server" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161779 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce87b9c-add6-4744-8463-0d268867b9de" containerName="barbican-keystone-listener-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161790 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ab3c4dd-5b70-4127-aa8d-570ffcde7477" containerName="nova-cell0-conductor-conductor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161800 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee97571-4ae5-4ed0-a32a-67abd5896ae0" containerName="cinder-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161813 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="container-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161823 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="account-reaper" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161839 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161849 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7b2888-eaed-4879-9f60-e4264aed2287" containerName="nova-api-log" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.161862 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ed394c7-5cf6-4a7b-8060-6ffb4953fbfd" containerName="ovn-controller" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162010 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="63999fb5-88e7-4df1-8084-267d0e37ac4c" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162104 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="73149d5c-c16f-4b83-b855-135fc1aab9fe" containerName="placement-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162139 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c1c989-3443-4b80-9786-56b68660655c" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162169 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96cfa1a-48d0-4efd-8594-4386a5d7ef35" containerName="openstack-network-exporter" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162198 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="01d907ea-d712-40af-b18a-5d55287fbaeb" containerName="mariadb-account-delete" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162218 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d677a57-1acf-4627-aa66-f4bee96a2b51" containerName="rabbitmq" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162247 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e84af371-9517-42aa-943b-a4847f4f8577" containerName="neutron-httpd" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162277 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ec7c62-702e-4794-81fc-64c72f72cb12" containerName="object-auditor" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.162303 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aedece6-346c-4003-8d00-95509a35f23b" containerName="barbican-api" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.163272 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.166207 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.170209 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.173499 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj"] Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.244547 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-config-volume\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.244617 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbpjs\" (UniqueName: \"kubernetes.io/projected/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-kube-api-access-jbpjs\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.244664 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-secret-volume\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.345945 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-config-volume\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.346270 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbpjs\" (UniqueName: \"kubernetes.io/projected/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-kube-api-access-jbpjs\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.346364 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-secret-volume\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.347468 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-config-volume\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.354004 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-secret-volume\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.363751 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbpjs\" (UniqueName: \"kubernetes.io/projected/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-kube-api-access-jbpjs\") pod \"collect-profiles-29398035-hh8jj\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.487834 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.939383 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj"] Nov 23 07:15:00 crc kubenswrapper[4906]: I1123 07:15:00.994475 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" event={"ID":"6fcb8821-fed8-4298-97c0-458ef9c6f6c2","Type":"ContainerStarted","Data":"48d46917d2b6c9b2818bf3ce5db86e883c3c41bb5eceaceae6472f642728ed92"} Nov 23 07:15:02 crc kubenswrapper[4906]: I1123 07:15:02.004644 4906 generic.go:334] "Generic (PLEG): container finished" podID="6fcb8821-fed8-4298-97c0-458ef9c6f6c2" containerID="2bec6b3863851e923a4cbe255526d460a7ce79bb7b309ab01a6cb0cb10cc6cac" exitCode=0 Nov 23 07:15:02 crc kubenswrapper[4906]: I1123 07:15:02.004724 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" event={"ID":"6fcb8821-fed8-4298-97c0-458ef9c6f6c2","Type":"ContainerDied","Data":"2bec6b3863851e923a4cbe255526d460a7ce79bb7b309ab01a6cb0cb10cc6cac"} Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.309656 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.394965 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbpjs\" (UniqueName: \"kubernetes.io/projected/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-kube-api-access-jbpjs\") pod \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.395101 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-secret-volume\") pod \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.395188 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-config-volume\") pod \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\" (UID: \"6fcb8821-fed8-4298-97c0-458ef9c6f6c2\") " Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.396638 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-config-volume" (OuterVolumeSpecName: "config-volume") pod "6fcb8821-fed8-4298-97c0-458ef9c6f6c2" (UID: "6fcb8821-fed8-4298-97c0-458ef9c6f6c2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.400803 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.404853 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6fcb8821-fed8-4298-97c0-458ef9c6f6c2" (UID: "6fcb8821-fed8-4298-97c0-458ef9c6f6c2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.414987 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-kube-api-access-jbpjs" (OuterVolumeSpecName: "kube-api-access-jbpjs") pod "6fcb8821-fed8-4298-97c0-458ef9c6f6c2" (UID: "6fcb8821-fed8-4298-97c0-458ef9c6f6c2"). InnerVolumeSpecName "kube-api-access-jbpjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.502103 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbpjs\" (UniqueName: \"kubernetes.io/projected/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-kube-api-access-jbpjs\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:03 crc kubenswrapper[4906]: I1123 07:15:03.502141 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fcb8821-fed8-4298-97c0-458ef9c6f6c2-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:04 crc kubenswrapper[4906]: I1123 07:15:04.024178 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" event={"ID":"6fcb8821-fed8-4298-97c0-458ef9c6f6c2","Type":"ContainerDied","Data":"48d46917d2b6c9b2818bf3ce5db86e883c3c41bb5eceaceae6472f642728ed92"} Nov 23 07:15:04 crc kubenswrapper[4906]: I1123 07:15:04.024639 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48d46917d2b6c9b2818bf3ce5db86e883c3c41bb5eceaceae6472f642728ed92" Nov 23 07:15:04 crc kubenswrapper[4906]: I1123 07:15:04.024749 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.289154 4906 scope.go:117] "RemoveContainer" containerID="eaab7475fcf8e8c5b70f5983da2df20065a5cb1d25322e150eb55c5ace653871" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.348033 4906 scope.go:117] "RemoveContainer" containerID="1964d4fe020c161bced8f66ff4439496702b1f867d376b389cd394821610ab26" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.400879 4906 scope.go:117] "RemoveContainer" containerID="feeab8377b8ce338f687bcf8f59f53e807ba5d4b1a8a47037bf6e81ca91565dc" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.442292 4906 scope.go:117] "RemoveContainer" containerID="d6b6def31b94f53e43ab11d1f9b18fccdd68fc5016706a678167f411d20a9456" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.462408 4906 scope.go:117] "RemoveContainer" containerID="018decda61999cccfd06ac4b20b5448abba52ae9229f9dd58c47095fdc6d2e57" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.487402 4906 scope.go:117] "RemoveContainer" containerID="64f8b0bc40e85813f6b53ebb1774deed7359fad207a70108c893d8f0a3d3af91" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.514427 4906 scope.go:117] "RemoveContainer" containerID="cbd45e5087e3ff34aceaedab7642e3f533b4b897abad75afa1f69b5f5f11f37f" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.550236 4906 scope.go:117] "RemoveContainer" containerID="a132f35e3274982037e61dc2cbbb3c263680061b23f75f55481fc4121df6fdc6" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.581507 4906 scope.go:117] "RemoveContainer" containerID="173d03c6a95edb199581f0d5117497414c7ba5718e997ef08b4ebdf4fb61cb7a" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.602090 4906 scope.go:117] "RemoveContainer" containerID="b7828b55bc0c0687cf95e500a55863b74ca5ad9cf605f0ed87a24da3adab43cb" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.627579 4906 scope.go:117] "RemoveContainer" containerID="64aa8715dc63cefdc0464b31e6dedda624baac63db9568706ac79df76a4f97a5" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.644930 4906 scope.go:117] "RemoveContainer" containerID="531081772b215724b14739dbe8db3d3198c4edaa7161148e8553a2bddb146101" Nov 23 07:15:54 crc kubenswrapper[4906]: I1123 07:15:54.676800 4906 scope.go:117] "RemoveContainer" containerID="739fbb582901e5e7181fb6829f18d4295a29898fbc690f867445b5383dea4f42" Nov 23 07:16:50 crc kubenswrapper[4906]: I1123 07:16:50.945655 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:16:50 crc kubenswrapper[4906]: I1123 07:16:50.946550 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:16:55 crc kubenswrapper[4906]: I1123 07:16:55.003860 4906 scope.go:117] "RemoveContainer" containerID="0dd9b19587494f220b9ce7cb80f137d8caf2f6b1707a1f5bae2b3bc3625d12af" Nov 23 07:16:55 crc kubenswrapper[4906]: I1123 07:16:55.036288 4906 scope.go:117] "RemoveContainer" containerID="e5fc6b77a408f81d5b91fd45ba27278acb2c5d247b602dbc2475870c5ba45b81" Nov 23 07:16:55 crc kubenswrapper[4906]: I1123 07:16:55.071313 4906 scope.go:117] "RemoveContainer" containerID="1f3a9d97d10c50b09e77dcba87dffe97001871bb27409ed0d6f948816f5a32ec" Nov 23 07:16:55 crc kubenswrapper[4906]: I1123 07:16:55.134272 4906 scope.go:117] "RemoveContainer" containerID="78e2be376a6ac18267e9fa24359c2479618c77eeaf00a6b4b20186ab50f161d6" Nov 23 07:16:55 crc kubenswrapper[4906]: I1123 07:16:55.159045 4906 scope.go:117] "RemoveContainer" containerID="a3add9b8cbef96dced17c8263087b9f52097f904de7cf1efc850f3e41c400acf" Nov 23 07:17:20 crc kubenswrapper[4906]: I1123 07:17:20.946407 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:17:20 crc kubenswrapper[4906]: I1123 07:17:20.947392 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:17:50 crc kubenswrapper[4906]: I1123 07:17:50.946921 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:17:50 crc kubenswrapper[4906]: I1123 07:17:50.951901 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:17:50 crc kubenswrapper[4906]: I1123 07:17:50.952019 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:17:50 crc kubenswrapper[4906]: I1123 07:17:50.953875 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:17:50 crc kubenswrapper[4906]: I1123 07:17:50.954033 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" gracePeriod=600 Nov 23 07:17:51 crc kubenswrapper[4906]: E1123 07:17:51.088502 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:17:51 crc kubenswrapper[4906]: I1123 07:17:51.965749 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" exitCode=0 Nov 23 07:17:51 crc kubenswrapper[4906]: I1123 07:17:51.966007 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7"} Nov 23 07:17:51 crc kubenswrapper[4906]: I1123 07:17:51.966295 4906 scope.go:117] "RemoveContainer" containerID="efb82b69034455635a0b7ce63f8c53d1bea8631681ec6e17e626273b28cdbdd5" Nov 23 07:17:51 crc kubenswrapper[4906]: I1123 07:17:51.967513 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:17:51 crc kubenswrapper[4906]: E1123 07:17:51.968029 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.250834 4906 scope.go:117] "RemoveContainer" containerID="dd3f69e2e62458ca440ee946a42dded4d158d99cfba6eb855675caa7f5eb42e5" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.291555 4906 scope.go:117] "RemoveContainer" containerID="85604e28eb73da42df364a73f2a404a0c2cac5adb3cdbd3da52bb12fba8bccc0" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.341124 4906 scope.go:117] "RemoveContainer" containerID="9fd4e4ce70e7b4e8a8a0cf7929ef4a4d7fd0ebe705d7489ea597f1eedf4b7d4e" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.374755 4906 scope.go:117] "RemoveContainer" containerID="6ba149ba0e302725b1fbc3ed83081a7cae1d95f6be1c61dedefc243562bcc3ec" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.405653 4906 scope.go:117] "RemoveContainer" containerID="6a774e7bd820a8715662dc5a9ea4bf2e91f3dcc4e667d067b8591e6ae75de44b" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.439795 4906 scope.go:117] "RemoveContainer" containerID="2c8ea7bf763beb872aa76646d1aa32281a56e75ac152c2352a663e16c3a97454" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.462099 4906 scope.go:117] "RemoveContainer" containerID="fb2db4868429914c4974cfcbe6924d07ccd95b04a3c7bb744e5c5081ac478bad" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.485295 4906 scope.go:117] "RemoveContainer" containerID="c76a5098fb2971db8ea7b34b60e4319599ca18452356b9c29903821f2d98dc07" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.512496 4906 scope.go:117] "RemoveContainer" containerID="2fd16719fcd506b038a0152208792ed142945112e8297ebc99ad99b173d43f11" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.532959 4906 scope.go:117] "RemoveContainer" containerID="659ee25bbd94087f3829983a3b02a6c010a9ff2d2e93cdd5a42d2537a3e03b42" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.559552 4906 scope.go:117] "RemoveContainer" containerID="a1ce0f6ba41a9ea922831c56c6043771d63ea9f48abecf10c7758bee8f51ec7e" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.601453 4906 scope.go:117] "RemoveContainer" containerID="aa8126291bff144f603418e592b1f6dcaccc4c065fbe8568ca377e65087939b3" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.625006 4906 scope.go:117] "RemoveContainer" containerID="f4c66a184ddf3fddd4f01f4bf5d9927e08a50644685ee93df83ca8a705617d12" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.651099 4906 scope.go:117] "RemoveContainer" containerID="9c4e48b39c14772f337cfc58ea53a79082fd8c022fe4b10170edefdfb0a64859" Nov 23 07:17:55 crc kubenswrapper[4906]: I1123 07:17:55.672517 4906 scope.go:117] "RemoveContainer" containerID="e67f16d70428889059a4cd8f993f06a74b72e1768492d989fa9a429a6fc5133b" Nov 23 07:18:03 crc kubenswrapper[4906]: I1123 07:18:03.356667 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:18:03 crc kubenswrapper[4906]: E1123 07:18:03.358071 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:18:17 crc kubenswrapper[4906]: I1123 07:18:17.357489 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:18:17 crc kubenswrapper[4906]: E1123 07:18:17.358885 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.132295 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k6ctt"] Nov 23 07:18:22 crc kubenswrapper[4906]: E1123 07:18:22.134038 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fcb8821-fed8-4298-97c0-458ef9c6f6c2" containerName="collect-profiles" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.134076 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fcb8821-fed8-4298-97c0-458ef9c6f6c2" containerName="collect-profiles" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.134554 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fcb8821-fed8-4298-97c0-458ef9c6f6c2" containerName="collect-profiles" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.137995 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.142111 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6ctt"] Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.299884 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgb9m\" (UniqueName: \"kubernetes.io/projected/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-kube-api-access-pgb9m\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.300083 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-utilities\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.300117 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-catalog-content\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.402203 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgb9m\" (UniqueName: \"kubernetes.io/projected/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-kube-api-access-pgb9m\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.402426 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-utilities\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.402462 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-catalog-content\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.403203 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-catalog-content\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.403428 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-utilities\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.434761 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgb9m\" (UniqueName: \"kubernetes.io/projected/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-kube-api-access-pgb9m\") pod \"certified-operators-k6ctt\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.468454 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:22 crc kubenswrapper[4906]: I1123 07:18:22.775054 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6ctt"] Nov 23 07:18:23 crc kubenswrapper[4906]: I1123 07:18:23.306127 4906 generic.go:334] "Generic (PLEG): container finished" podID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerID="bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212" exitCode=0 Nov 23 07:18:23 crc kubenswrapper[4906]: I1123 07:18:23.306346 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ctt" event={"ID":"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8","Type":"ContainerDied","Data":"bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212"} Nov 23 07:18:23 crc kubenswrapper[4906]: I1123 07:18:23.306534 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ctt" event={"ID":"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8","Type":"ContainerStarted","Data":"562951d1f1ed707062894570f51ef7d6613a64146ce87da1dc5a5ab0daa88333"} Nov 23 07:18:23 crc kubenswrapper[4906]: I1123 07:18:23.310056 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:18:24 crc kubenswrapper[4906]: I1123 07:18:24.325369 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ctt" event={"ID":"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8","Type":"ContainerStarted","Data":"d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4"} Nov 23 07:18:25 crc kubenswrapper[4906]: I1123 07:18:25.341193 4906 generic.go:334] "Generic (PLEG): container finished" podID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerID="d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4" exitCode=0 Nov 23 07:18:25 crc kubenswrapper[4906]: I1123 07:18:25.341290 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ctt" event={"ID":"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8","Type":"ContainerDied","Data":"d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4"} Nov 23 07:18:26 crc kubenswrapper[4906]: I1123 07:18:26.354113 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ctt" event={"ID":"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8","Type":"ContainerStarted","Data":"a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb"} Nov 23 07:18:26 crc kubenswrapper[4906]: I1123 07:18:26.389514 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k6ctt" podStartSLOduration=1.897087985 podStartE2EDuration="4.389491665s" podCreationTimestamp="2025-11-23 07:18:22 +0000 UTC" firstStartedPulling="2025-11-23 07:18:23.309566248 +0000 UTC m=+1718.822957581" lastFinishedPulling="2025-11-23 07:18:25.801969948 +0000 UTC m=+1721.315361261" observedRunningTime="2025-11-23 07:18:26.387307717 +0000 UTC m=+1721.900699020" watchObservedRunningTime="2025-11-23 07:18:26.389491665 +0000 UTC m=+1721.902882968" Nov 23 07:18:28 crc kubenswrapper[4906]: I1123 07:18:28.356591 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:18:28 crc kubenswrapper[4906]: E1123 07:18:28.357233 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:18:32 crc kubenswrapper[4906]: I1123 07:18:32.468802 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:32 crc kubenswrapper[4906]: I1123 07:18:32.469483 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:32 crc kubenswrapper[4906]: I1123 07:18:32.557206 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:33 crc kubenswrapper[4906]: I1123 07:18:33.528775 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:39 crc kubenswrapper[4906]: I1123 07:18:39.707167 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6ctt"] Nov 23 07:18:39 crc kubenswrapper[4906]: I1123 07:18:39.708462 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k6ctt" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="registry-server" containerID="cri-o://a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb" gracePeriod=2 Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.260391 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.358237 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:18:40 crc kubenswrapper[4906]: E1123 07:18:40.358803 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.423185 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-utilities\") pod \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.423425 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgb9m\" (UniqueName: \"kubernetes.io/projected/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-kube-api-access-pgb9m\") pod \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.423510 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-catalog-content\") pod \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\" (UID: \"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8\") " Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.425183 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-utilities" (OuterVolumeSpecName: "utilities") pod "85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" (UID: "85dadc3c-e560-4a00-8075-8dbf3cf6d5a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.434013 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-kube-api-access-pgb9m" (OuterVolumeSpecName: "kube-api-access-pgb9m") pod "85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" (UID: "85dadc3c-e560-4a00-8075-8dbf3cf6d5a8"). InnerVolumeSpecName "kube-api-access-pgb9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.512381 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" (UID: "85dadc3c-e560-4a00-8075-8dbf3cf6d5a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.526173 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgb9m\" (UniqueName: \"kubernetes.io/projected/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-kube-api-access-pgb9m\") on node \"crc\" DevicePath \"\"" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.526209 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.526224 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.549437 4906 generic.go:334] "Generic (PLEG): container finished" podID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerID="a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb" exitCode=0 Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.549544 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6ctt" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.549579 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ctt" event={"ID":"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8","Type":"ContainerDied","Data":"a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb"} Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.549783 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6ctt" event={"ID":"85dadc3c-e560-4a00-8075-8dbf3cf6d5a8","Type":"ContainerDied","Data":"562951d1f1ed707062894570f51ef7d6613a64146ce87da1dc5a5ab0daa88333"} Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.549871 4906 scope.go:117] "RemoveContainer" containerID="a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.594488 4906 scope.go:117] "RemoveContainer" containerID="d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.608827 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6ctt"] Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.616625 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k6ctt"] Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.635373 4906 scope.go:117] "RemoveContainer" containerID="bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.685726 4906 scope.go:117] "RemoveContainer" containerID="a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb" Nov 23 07:18:40 crc kubenswrapper[4906]: E1123 07:18:40.686495 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb\": container with ID starting with a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb not found: ID does not exist" containerID="a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.686575 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb"} err="failed to get container status \"a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb\": rpc error: code = NotFound desc = could not find container \"a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb\": container with ID starting with a6db3b5ff870701107be22f02ef4a74542b9aa91ab8d7074d497e0d9b3896ebb not found: ID does not exist" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.686635 4906 scope.go:117] "RemoveContainer" containerID="d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4" Nov 23 07:18:40 crc kubenswrapper[4906]: E1123 07:18:40.687388 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4\": container with ID starting with d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4 not found: ID does not exist" containerID="d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.687434 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4"} err="failed to get container status \"d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4\": rpc error: code = NotFound desc = could not find container \"d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4\": container with ID starting with d89466fc25ce01fb8813b5f6ab85b1a99f7f05a5ac6f5de383d134d961475fb4 not found: ID does not exist" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.687466 4906 scope.go:117] "RemoveContainer" containerID="bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212" Nov 23 07:18:40 crc kubenswrapper[4906]: E1123 07:18:40.688402 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212\": container with ID starting with bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212 not found: ID does not exist" containerID="bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212" Nov 23 07:18:40 crc kubenswrapper[4906]: I1123 07:18:40.688452 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212"} err="failed to get container status \"bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212\": rpc error: code = NotFound desc = could not find container \"bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212\": container with ID starting with bf41fd1a685bfa462b7fa2c9fb2861bb56ce10edf6c09a4ccc8279affa019212 not found: ID does not exist" Nov 23 07:18:41 crc kubenswrapper[4906]: I1123 07:18:41.369201 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" path="/var/lib/kubelet/pods/85dadc3c-e560-4a00-8075-8dbf3cf6d5a8/volumes" Nov 23 07:18:55 crc kubenswrapper[4906]: I1123 07:18:55.361922 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:18:55 crc kubenswrapper[4906]: E1123 07:18:55.363315 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:18:55 crc kubenswrapper[4906]: I1123 07:18:55.840717 4906 scope.go:117] "RemoveContainer" containerID="aee41b6a5dd1ef9d39f6a37ecd59b25e04cf324a7f6135066b76c4a4552915f9" Nov 23 07:18:55 crc kubenswrapper[4906]: I1123 07:18:55.866441 4906 scope.go:117] "RemoveContainer" containerID="6f1bb39f61ceebf4bbb9cb405575f82b1dff5097833cdf28d70e297480072af0" Nov 23 07:18:55 crc kubenswrapper[4906]: I1123 07:18:55.970435 4906 scope.go:117] "RemoveContainer" containerID="9cf15debcaa784ebb7c366d5efc1dfc07119d5f901f8b3e1db3f4dfeef24c333" Nov 23 07:18:55 crc kubenswrapper[4906]: I1123 07:18:55.998221 4906 scope.go:117] "RemoveContainer" containerID="f1bcaee5f685e06a5e0f095d05465381185a3d0e78e940373d9a346353454dc6" Nov 23 07:18:56 crc kubenswrapper[4906]: I1123 07:18:56.044416 4906 scope.go:117] "RemoveContainer" containerID="96435a9a6698e54650b138b7b8daa6f2fcc2fc3938e89653a935fea62eeb5cf6" Nov 23 07:19:08 crc kubenswrapper[4906]: I1123 07:19:08.361040 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:19:08 crc kubenswrapper[4906]: E1123 07:19:08.363214 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:19:21 crc kubenswrapper[4906]: I1123 07:19:21.357358 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:19:21 crc kubenswrapper[4906]: E1123 07:19:21.359260 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:19:33 crc kubenswrapper[4906]: I1123 07:19:33.356933 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:19:33 crc kubenswrapper[4906]: E1123 07:19:33.358176 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:19:48 crc kubenswrapper[4906]: I1123 07:19:48.356636 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:19:48 crc kubenswrapper[4906]: E1123 07:19:48.358254 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:19:56 crc kubenswrapper[4906]: I1123 07:19:56.178029 4906 scope.go:117] "RemoveContainer" containerID="8da666a9e40465436d1857a8452785646b2c0c473e10ad242db4541b87e6a037" Nov 23 07:19:56 crc kubenswrapper[4906]: I1123 07:19:56.208088 4906 scope.go:117] "RemoveContainer" containerID="abee71bb44307a2c4573a7acf7731e1011f5ab2e46aced3873ac6dfb10570fe3" Nov 23 07:19:56 crc kubenswrapper[4906]: I1123 07:19:56.233573 4906 scope.go:117] "RemoveContainer" containerID="4390d57fff98f39e182633e92a5feb33d5eab146a05e1099cbd32ea1b43c437f" Nov 23 07:19:56 crc kubenswrapper[4906]: I1123 07:19:56.257377 4906 scope.go:117] "RemoveContainer" containerID="cfe0d8e3ca16e113f2016b61cb50f6b5b415f288a07bd395effcd5f6002c7ed6" Nov 23 07:19:56 crc kubenswrapper[4906]: I1123 07:19:56.277192 4906 scope.go:117] "RemoveContainer" containerID="f7be6a395372de3ef443ddccc155ec0954c9663ac6663a6358c697025786322a" Nov 23 07:19:56 crc kubenswrapper[4906]: I1123 07:19:56.292943 4906 scope.go:117] "RemoveContainer" containerID="59399848d4fd8c548b0ad9fe61db4277a707a18507241aca88683dc29eebb13a" Nov 23 07:19:56 crc kubenswrapper[4906]: I1123 07:19:56.327060 4906 scope.go:117] "RemoveContainer" containerID="168ec284586fb5aa74b2d96f311c089e82fd4a4806a8147df56bdd9a2c39ee76" Nov 23 07:20:01 crc kubenswrapper[4906]: I1123 07:20:01.357229 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:20:01 crc kubenswrapper[4906]: E1123 07:20:01.358010 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:20:03 crc kubenswrapper[4906]: I1123 07:20:03.369956 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" podUID="3bc07001-7205-48f6-a348-e62df3c098da" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.74:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:20:03 crc kubenswrapper[4906]: I1123 07:20:03.370254 4906 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-bgshg" podUID="3bc07001-7205-48f6-a348-e62df3c098da" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.74:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:20:16 crc kubenswrapper[4906]: I1123 07:20:16.356844 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:20:16 crc kubenswrapper[4906]: E1123 07:20:16.358176 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:20:27 crc kubenswrapper[4906]: I1123 07:20:27.357495 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:20:27 crc kubenswrapper[4906]: E1123 07:20:27.358701 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:20:38 crc kubenswrapper[4906]: I1123 07:20:38.357413 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:20:38 crc kubenswrapper[4906]: E1123 07:20:38.358957 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:20:51 crc kubenswrapper[4906]: I1123 07:20:51.357510 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:20:51 crc kubenswrapper[4906]: E1123 07:20:51.358872 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:21:06 crc kubenswrapper[4906]: I1123 07:21:06.357506 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:21:06 crc kubenswrapper[4906]: E1123 07:21:06.359289 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:21:20 crc kubenswrapper[4906]: I1123 07:21:20.357754 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:21:20 crc kubenswrapper[4906]: E1123 07:21:20.359314 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:21:31 crc kubenswrapper[4906]: I1123 07:21:31.357426 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:21:31 crc kubenswrapper[4906]: E1123 07:21:31.358964 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:21:44 crc kubenswrapper[4906]: I1123 07:21:44.356259 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:21:44 crc kubenswrapper[4906]: E1123 07:21:44.357394 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:21:59 crc kubenswrapper[4906]: I1123 07:21:59.356582 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:21:59 crc kubenswrapper[4906]: E1123 07:21:59.357463 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:22:14 crc kubenswrapper[4906]: I1123 07:22:14.356557 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:22:14 crc kubenswrapper[4906]: E1123 07:22:14.357973 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:22:27 crc kubenswrapper[4906]: I1123 07:22:27.357770 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:22:27 crc kubenswrapper[4906]: E1123 07:22:27.361131 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:22:40 crc kubenswrapper[4906]: I1123 07:22:40.356988 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:22:40 crc kubenswrapper[4906]: E1123 07:22:40.358218 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:22:53 crc kubenswrapper[4906]: I1123 07:22:53.356835 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:22:54 crc kubenswrapper[4906]: I1123 07:22:54.206429 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"7e38ec44b499cd9cd90308d44b4f055bb77948119e4077d14978ca4ea613b870"} Nov 23 07:23:02 crc kubenswrapper[4906]: I1123 07:23:02.919916 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9dtp4"] Nov 23 07:23:02 crc kubenswrapper[4906]: E1123 07:23:02.926414 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="extract-content" Nov 23 07:23:02 crc kubenswrapper[4906]: I1123 07:23:02.926657 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="extract-content" Nov 23 07:23:02 crc kubenswrapper[4906]: E1123 07:23:02.926774 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="registry-server" Nov 23 07:23:02 crc kubenswrapper[4906]: I1123 07:23:02.926891 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="registry-server" Nov 23 07:23:02 crc kubenswrapper[4906]: E1123 07:23:02.926990 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="extract-utilities" Nov 23 07:23:02 crc kubenswrapper[4906]: I1123 07:23:02.927067 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="extract-utilities" Nov 23 07:23:02 crc kubenswrapper[4906]: I1123 07:23:02.927332 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="85dadc3c-e560-4a00-8075-8dbf3cf6d5a8" containerName="registry-server" Nov 23 07:23:02 crc kubenswrapper[4906]: I1123 07:23:02.931115 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:02 crc kubenswrapper[4906]: I1123 07:23:02.931309 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dtp4"] Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.075012 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-utilities\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.075452 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-catalog-content\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.075560 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw7vt\" (UniqueName: \"kubernetes.io/projected/5b3b9742-ca25-4a3f-a362-25e8e414fc64-kube-api-access-dw7vt\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.177342 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-utilities\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.177410 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-catalog-content\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.177445 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw7vt\" (UniqueName: \"kubernetes.io/projected/5b3b9742-ca25-4a3f-a362-25e8e414fc64-kube-api-access-dw7vt\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.178148 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-utilities\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.178364 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-catalog-content\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.197326 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw7vt\" (UniqueName: \"kubernetes.io/projected/5b3b9742-ca25-4a3f-a362-25e8e414fc64-kube-api-access-dw7vt\") pod \"redhat-operators-9dtp4\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.298676 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:03 crc kubenswrapper[4906]: I1123 07:23:03.529153 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dtp4"] Nov 23 07:23:04 crc kubenswrapper[4906]: I1123 07:23:04.314653 4906 generic.go:334] "Generic (PLEG): container finished" podID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerID="f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581" exitCode=0 Nov 23 07:23:04 crc kubenswrapper[4906]: I1123 07:23:04.314787 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dtp4" event={"ID":"5b3b9742-ca25-4a3f-a362-25e8e414fc64","Type":"ContainerDied","Data":"f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581"} Nov 23 07:23:04 crc kubenswrapper[4906]: I1123 07:23:04.315152 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dtp4" event={"ID":"5b3b9742-ca25-4a3f-a362-25e8e414fc64","Type":"ContainerStarted","Data":"cc036d034d4b269b63a89ebd0d1eaf00f50544075f6ef513d9d7a83313a44198"} Nov 23 07:23:05 crc kubenswrapper[4906]: I1123 07:23:05.328200 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dtp4" event={"ID":"5b3b9742-ca25-4a3f-a362-25e8e414fc64","Type":"ContainerStarted","Data":"74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa"} Nov 23 07:23:06 crc kubenswrapper[4906]: I1123 07:23:06.342004 4906 generic.go:334] "Generic (PLEG): container finished" podID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerID="74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa" exitCode=0 Nov 23 07:23:06 crc kubenswrapper[4906]: I1123 07:23:06.342091 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dtp4" event={"ID":"5b3b9742-ca25-4a3f-a362-25e8e414fc64","Type":"ContainerDied","Data":"74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa"} Nov 23 07:23:07 crc kubenswrapper[4906]: I1123 07:23:07.386105 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dtp4" event={"ID":"5b3b9742-ca25-4a3f-a362-25e8e414fc64","Type":"ContainerStarted","Data":"92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf"} Nov 23 07:23:13 crc kubenswrapper[4906]: I1123 07:23:13.299340 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:13 crc kubenswrapper[4906]: I1123 07:23:13.300643 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:14 crc kubenswrapper[4906]: I1123 07:23:14.373904 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9dtp4" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="registry-server" probeResult="failure" output=< Nov 23 07:23:14 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 07:23:14 crc kubenswrapper[4906]: > Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.046038 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9dtp4" podStartSLOduration=15.604913224 podStartE2EDuration="18.046001554s" podCreationTimestamp="2025-11-23 07:23:02 +0000 UTC" firstStartedPulling="2025-11-23 07:23:04.316583528 +0000 UTC m=+1999.829974831" lastFinishedPulling="2025-11-23 07:23:06.757671798 +0000 UTC m=+2002.271063161" observedRunningTime="2025-11-23 07:23:07.385391607 +0000 UTC m=+2002.898782920" watchObservedRunningTime="2025-11-23 07:23:20.046001554 +0000 UTC m=+2015.559392897" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.047174 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vbrlg"] Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.049974 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.076176 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbrlg"] Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.216066 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-catalog-content\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.216368 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k957j\" (UniqueName: \"kubernetes.io/projected/e515e045-a78e-46cb-849a-006d730dfdd3-kube-api-access-k957j\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.216513 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-utilities\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.317995 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k957j\" (UniqueName: \"kubernetes.io/projected/e515e045-a78e-46cb-849a-006d730dfdd3-kube-api-access-k957j\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.318097 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-utilities\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.318219 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-catalog-content\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.318713 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-utilities\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.318915 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-catalog-content\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.352822 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k957j\" (UniqueName: \"kubernetes.io/projected/e515e045-a78e-46cb-849a-006d730dfdd3-kube-api-access-k957j\") pod \"redhat-marketplace-vbrlg\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.373213 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:20 crc kubenswrapper[4906]: I1123 07:23:20.840430 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbrlg"] Nov 23 07:23:21 crc kubenswrapper[4906]: I1123 07:23:21.542886 4906 generic.go:334] "Generic (PLEG): container finished" podID="e515e045-a78e-46cb-849a-006d730dfdd3" containerID="1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887" exitCode=0 Nov 23 07:23:21 crc kubenswrapper[4906]: I1123 07:23:21.542988 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbrlg" event={"ID":"e515e045-a78e-46cb-849a-006d730dfdd3","Type":"ContainerDied","Data":"1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887"} Nov 23 07:23:21 crc kubenswrapper[4906]: I1123 07:23:21.543050 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbrlg" event={"ID":"e515e045-a78e-46cb-849a-006d730dfdd3","Type":"ContainerStarted","Data":"a446cc7ec1b4e85c29158dc34c6b1d718173348c95ea384e705d8ba94d69279a"} Nov 23 07:23:23 crc kubenswrapper[4906]: I1123 07:23:23.371256 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:23 crc kubenswrapper[4906]: I1123 07:23:23.434066 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:23 crc kubenswrapper[4906]: I1123 07:23:23.567902 4906 generic.go:334] "Generic (PLEG): container finished" podID="e515e045-a78e-46cb-849a-006d730dfdd3" containerID="4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090" exitCode=0 Nov 23 07:23:23 crc kubenswrapper[4906]: I1123 07:23:23.568103 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbrlg" event={"ID":"e515e045-a78e-46cb-849a-006d730dfdd3","Type":"ContainerDied","Data":"4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090"} Nov 23 07:23:23 crc kubenswrapper[4906]: I1123 07:23:23.571476 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:23:24 crc kubenswrapper[4906]: I1123 07:23:24.582601 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbrlg" event={"ID":"e515e045-a78e-46cb-849a-006d730dfdd3","Type":"ContainerStarted","Data":"ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc"} Nov 23 07:23:24 crc kubenswrapper[4906]: I1123 07:23:24.604806 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vbrlg" podStartSLOduration=2.166532283 podStartE2EDuration="4.604676323s" podCreationTimestamp="2025-11-23 07:23:20 +0000 UTC" firstStartedPulling="2025-11-23 07:23:21.545747176 +0000 UTC m=+2017.059138509" lastFinishedPulling="2025-11-23 07:23:23.983891246 +0000 UTC m=+2019.497282549" observedRunningTime="2025-11-23 07:23:24.600328388 +0000 UTC m=+2020.113719701" watchObservedRunningTime="2025-11-23 07:23:24.604676323 +0000 UTC m=+2020.118067636" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.008072 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dtp4"] Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.008386 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9dtp4" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="registry-server" containerID="cri-o://92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf" gracePeriod=2 Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.482018 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.596784 4906 generic.go:334] "Generic (PLEG): container finished" podID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerID="92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf" exitCode=0 Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.597809 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dtp4" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.598266 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dtp4" event={"ID":"5b3b9742-ca25-4a3f-a362-25e8e414fc64","Type":"ContainerDied","Data":"92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf"} Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.598295 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dtp4" event={"ID":"5b3b9742-ca25-4a3f-a362-25e8e414fc64","Type":"ContainerDied","Data":"cc036d034d4b269b63a89ebd0d1eaf00f50544075f6ef513d9d7a83313a44198"} Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.598313 4906 scope.go:117] "RemoveContainer" containerID="92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.623636 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-utilities\") pod \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.623682 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw7vt\" (UniqueName: \"kubernetes.io/projected/5b3b9742-ca25-4a3f-a362-25e8e414fc64-kube-api-access-dw7vt\") pod \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.623796 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-catalog-content\") pod \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\" (UID: \"5b3b9742-ca25-4a3f-a362-25e8e414fc64\") " Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.624892 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-utilities" (OuterVolumeSpecName: "utilities") pod "5b3b9742-ca25-4a3f-a362-25e8e414fc64" (UID: "5b3b9742-ca25-4a3f-a362-25e8e414fc64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.629985 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b3b9742-ca25-4a3f-a362-25e8e414fc64-kube-api-access-dw7vt" (OuterVolumeSpecName: "kube-api-access-dw7vt") pod "5b3b9742-ca25-4a3f-a362-25e8e414fc64" (UID: "5b3b9742-ca25-4a3f-a362-25e8e414fc64"). InnerVolumeSpecName "kube-api-access-dw7vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.633589 4906 scope.go:117] "RemoveContainer" containerID="74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.687693 4906 scope.go:117] "RemoveContainer" containerID="f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.713327 4906 scope.go:117] "RemoveContainer" containerID="92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf" Nov 23 07:23:25 crc kubenswrapper[4906]: E1123 07:23:25.713966 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf\": container with ID starting with 92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf not found: ID does not exist" containerID="92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.714013 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf"} err="failed to get container status \"92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf\": rpc error: code = NotFound desc = could not find container \"92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf\": container with ID starting with 92730d1e1ed2d1fbd5a0be7f14f8e82d12f7a9f1347c92e05c7bf48a123be6bf not found: ID does not exist" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.714049 4906 scope.go:117] "RemoveContainer" containerID="74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa" Nov 23 07:23:25 crc kubenswrapper[4906]: E1123 07:23:25.714566 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa\": container with ID starting with 74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa not found: ID does not exist" containerID="74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.714627 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa"} err="failed to get container status \"74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa\": rpc error: code = NotFound desc = could not find container \"74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa\": container with ID starting with 74d860c5f186ea1aa1ecdb476b40d02f6cfcc38530f602f01fc16faf01ea40aa not found: ID does not exist" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.714648 4906 scope.go:117] "RemoveContainer" containerID="f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581" Nov 23 07:23:25 crc kubenswrapper[4906]: E1123 07:23:25.715086 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581\": container with ID starting with f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581 not found: ID does not exist" containerID="f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.715119 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581"} err="failed to get container status \"f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581\": rpc error: code = NotFound desc = could not find container \"f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581\": container with ID starting with f08c3ce40bcce9a34ea9dba3b7d5aec74c1fc5f04e69ad57e613fd05b75f3581 not found: ID does not exist" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.725337 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.725361 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw7vt\" (UniqueName: \"kubernetes.io/projected/5b3b9742-ca25-4a3f-a362-25e8e414fc64-kube-api-access-dw7vt\") on node \"crc\" DevicePath \"\"" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.745210 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b3b9742-ca25-4a3f-a362-25e8e414fc64" (UID: "5b3b9742-ca25-4a3f-a362-25e8e414fc64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.826576 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3b9742-ca25-4a3f-a362-25e8e414fc64-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.949198 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dtp4"] Nov 23 07:23:25 crc kubenswrapper[4906]: I1123 07:23:25.958714 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9dtp4"] Nov 23 07:23:27 crc kubenswrapper[4906]: I1123 07:23:27.370926 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" path="/var/lib/kubelet/pods/5b3b9742-ca25-4a3f-a362-25e8e414fc64/volumes" Nov 23 07:23:30 crc kubenswrapper[4906]: I1123 07:23:30.373819 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:30 crc kubenswrapper[4906]: I1123 07:23:30.375587 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:30 crc kubenswrapper[4906]: I1123 07:23:30.413802 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:30 crc kubenswrapper[4906]: I1123 07:23:30.704117 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:30 crc kubenswrapper[4906]: I1123 07:23:30.772436 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbrlg"] Nov 23 07:23:32 crc kubenswrapper[4906]: I1123 07:23:32.664049 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vbrlg" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="registry-server" containerID="cri-o://ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc" gracePeriod=2 Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.063980 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.155751 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-catalog-content\") pod \"e515e045-a78e-46cb-849a-006d730dfdd3\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.155912 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-utilities\") pod \"e515e045-a78e-46cb-849a-006d730dfdd3\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.155957 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k957j\" (UniqueName: \"kubernetes.io/projected/e515e045-a78e-46cb-849a-006d730dfdd3-kube-api-access-k957j\") pod \"e515e045-a78e-46cb-849a-006d730dfdd3\" (UID: \"e515e045-a78e-46cb-849a-006d730dfdd3\") " Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.157408 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-utilities" (OuterVolumeSpecName: "utilities") pod "e515e045-a78e-46cb-849a-006d730dfdd3" (UID: "e515e045-a78e-46cb-849a-006d730dfdd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.161822 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e515e045-a78e-46cb-849a-006d730dfdd3-kube-api-access-k957j" (OuterVolumeSpecName: "kube-api-access-k957j") pod "e515e045-a78e-46cb-849a-006d730dfdd3" (UID: "e515e045-a78e-46cb-849a-006d730dfdd3"). InnerVolumeSpecName "kube-api-access-k957j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.173835 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e515e045-a78e-46cb-849a-006d730dfdd3" (UID: "e515e045-a78e-46cb-849a-006d730dfdd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.257892 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.257958 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k957j\" (UniqueName: \"kubernetes.io/projected/e515e045-a78e-46cb-849a-006d730dfdd3-kube-api-access-k957j\") on node \"crc\" DevicePath \"\"" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.257979 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e515e045-a78e-46cb-849a-006d730dfdd3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.675097 4906 generic.go:334] "Generic (PLEG): container finished" podID="e515e045-a78e-46cb-849a-006d730dfdd3" containerID="ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc" exitCode=0 Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.675179 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vbrlg" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.675193 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbrlg" event={"ID":"e515e045-a78e-46cb-849a-006d730dfdd3","Type":"ContainerDied","Data":"ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc"} Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.676354 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vbrlg" event={"ID":"e515e045-a78e-46cb-849a-006d730dfdd3","Type":"ContainerDied","Data":"a446cc7ec1b4e85c29158dc34c6b1d718173348c95ea384e705d8ba94d69279a"} Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.676384 4906 scope.go:117] "RemoveContainer" containerID="ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.703157 4906 scope.go:117] "RemoveContainer" containerID="4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.706941 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbrlg"] Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.713248 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vbrlg"] Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.726993 4906 scope.go:117] "RemoveContainer" containerID="1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.777592 4906 scope.go:117] "RemoveContainer" containerID="ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc" Nov 23 07:23:33 crc kubenswrapper[4906]: E1123 07:23:33.778202 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc\": container with ID starting with ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc not found: ID does not exist" containerID="ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.778242 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc"} err="failed to get container status \"ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc\": rpc error: code = NotFound desc = could not find container \"ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc\": container with ID starting with ebe5d87f05938047cf3cf71556a2759ef0296421a2d70fdac01717d950c30dbc not found: ID does not exist" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.778271 4906 scope.go:117] "RemoveContainer" containerID="4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090" Nov 23 07:23:33 crc kubenswrapper[4906]: E1123 07:23:33.778815 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090\": container with ID starting with 4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090 not found: ID does not exist" containerID="4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.778850 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090"} err="failed to get container status \"4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090\": rpc error: code = NotFound desc = could not find container \"4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090\": container with ID starting with 4a3dfea5fddefae7fc0ea9fdd25c137dedee5a5dd5d4d6eb4dad0473bba92090 not found: ID does not exist" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.778891 4906 scope.go:117] "RemoveContainer" containerID="1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887" Nov 23 07:23:33 crc kubenswrapper[4906]: E1123 07:23:33.779323 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887\": container with ID starting with 1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887 not found: ID does not exist" containerID="1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887" Nov 23 07:23:33 crc kubenswrapper[4906]: I1123 07:23:33.779401 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887"} err="failed to get container status \"1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887\": rpc error: code = NotFound desc = could not find container \"1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887\": container with ID starting with 1b2ae6420cbd951cadc3290fb67e86b8a87639427619afa4cbc3f1cda305b887 not found: ID does not exist" Nov 23 07:23:35 crc kubenswrapper[4906]: I1123 07:23:35.368057 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" path="/var/lib/kubelet/pods/e515e045-a78e-46cb-849a-006d730dfdd3/volumes" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.606227 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jgm8r"] Nov 23 07:24:18 crc kubenswrapper[4906]: E1123 07:24:18.607336 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="extract-utilities" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607353 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="extract-utilities" Nov 23 07:24:18 crc kubenswrapper[4906]: E1123 07:24:18.607366 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="registry-server" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607372 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="registry-server" Nov 23 07:24:18 crc kubenswrapper[4906]: E1123 07:24:18.607380 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="registry-server" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607387 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="registry-server" Nov 23 07:24:18 crc kubenswrapper[4906]: E1123 07:24:18.607407 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="extract-utilities" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607414 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="extract-utilities" Nov 23 07:24:18 crc kubenswrapper[4906]: E1123 07:24:18.607425 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="extract-content" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607430 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="extract-content" Nov 23 07:24:18 crc kubenswrapper[4906]: E1123 07:24:18.607445 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="extract-content" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607451 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="extract-content" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607590 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3b9742-ca25-4a3f-a362-25e8e414fc64" containerName="registry-server" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.607603 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e515e045-a78e-46cb-849a-006d730dfdd3" containerName="registry-server" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.608900 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.626282 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jgm8r"] Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.790525 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqgzs\" (UniqueName: \"kubernetes.io/projected/eecc9e0d-b884-489e-a67d-b4cf63f931c1-kube-api-access-tqgzs\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.790897 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eecc9e0d-b884-489e-a67d-b4cf63f931c1-catalog-content\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.790939 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eecc9e0d-b884-489e-a67d-b4cf63f931c1-utilities\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.892221 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eecc9e0d-b884-489e-a67d-b4cf63f931c1-catalog-content\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.892315 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eecc9e0d-b884-489e-a67d-b4cf63f931c1-utilities\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.892376 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqgzs\" (UniqueName: \"kubernetes.io/projected/eecc9e0d-b884-489e-a67d-b4cf63f931c1-kube-api-access-tqgzs\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.892909 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eecc9e0d-b884-489e-a67d-b4cf63f931c1-catalog-content\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.892944 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eecc9e0d-b884-489e-a67d-b4cf63f931c1-utilities\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.914220 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqgzs\" (UniqueName: \"kubernetes.io/projected/eecc9e0d-b884-489e-a67d-b4cf63f931c1-kube-api-access-tqgzs\") pod \"community-operators-jgm8r\" (UID: \"eecc9e0d-b884-489e-a67d-b4cf63f931c1\") " pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:18 crc kubenswrapper[4906]: I1123 07:24:18.940090 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:19 crc kubenswrapper[4906]: I1123 07:24:19.209514 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jgm8r"] Nov 23 07:24:20 crc kubenswrapper[4906]: I1123 07:24:20.162576 4906 generic.go:334] "Generic (PLEG): container finished" podID="eecc9e0d-b884-489e-a67d-b4cf63f931c1" containerID="a641c1a9e594696506ce85e66e465c07beefea6605e86099d3be59687f161181" exitCode=0 Nov 23 07:24:20 crc kubenswrapper[4906]: I1123 07:24:20.162626 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgm8r" event={"ID":"eecc9e0d-b884-489e-a67d-b4cf63f931c1","Type":"ContainerDied","Data":"a641c1a9e594696506ce85e66e465c07beefea6605e86099d3be59687f161181"} Nov 23 07:24:20 crc kubenswrapper[4906]: I1123 07:24:20.162655 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgm8r" event={"ID":"eecc9e0d-b884-489e-a67d-b4cf63f931c1","Type":"ContainerStarted","Data":"55f531a73e9710d12bea806c73af62f71936f1f0aaf24fbd93713d10f80eef3f"} Nov 23 07:24:24 crc kubenswrapper[4906]: I1123 07:24:24.211358 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgm8r" event={"ID":"eecc9e0d-b884-489e-a67d-b4cf63f931c1","Type":"ContainerStarted","Data":"0c010cef5861aef7d7f2cfe09f4332cc04314b8cdc802cd50bf9617c64ffe6cf"} Nov 23 07:24:25 crc kubenswrapper[4906]: I1123 07:24:25.223104 4906 generic.go:334] "Generic (PLEG): container finished" podID="eecc9e0d-b884-489e-a67d-b4cf63f931c1" containerID="0c010cef5861aef7d7f2cfe09f4332cc04314b8cdc802cd50bf9617c64ffe6cf" exitCode=0 Nov 23 07:24:25 crc kubenswrapper[4906]: I1123 07:24:25.223165 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgm8r" event={"ID":"eecc9e0d-b884-489e-a67d-b4cf63f931c1","Type":"ContainerDied","Data":"0c010cef5861aef7d7f2cfe09f4332cc04314b8cdc802cd50bf9617c64ffe6cf"} Nov 23 07:24:26 crc kubenswrapper[4906]: I1123 07:24:26.236027 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgm8r" event={"ID":"eecc9e0d-b884-489e-a67d-b4cf63f931c1","Type":"ContainerStarted","Data":"f544ac0500ff3377de30696bee31ab930d0e2a358dc259ea4e588ace1c75837e"} Nov 23 07:24:26 crc kubenswrapper[4906]: I1123 07:24:26.267777 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jgm8r" podStartSLOduration=2.845573029 podStartE2EDuration="8.267751692s" podCreationTimestamp="2025-11-23 07:24:18 +0000 UTC" firstStartedPulling="2025-11-23 07:24:20.164831334 +0000 UTC m=+2075.678222637" lastFinishedPulling="2025-11-23 07:24:25.587009997 +0000 UTC m=+2081.100401300" observedRunningTime="2025-11-23 07:24:26.261642861 +0000 UTC m=+2081.775034174" watchObservedRunningTime="2025-11-23 07:24:26.267751692 +0000 UTC m=+2081.781143005" Nov 23 07:24:28 crc kubenswrapper[4906]: I1123 07:24:28.940876 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:28 crc kubenswrapper[4906]: I1123 07:24:28.941349 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:29 crc kubenswrapper[4906]: I1123 07:24:29.023899 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.006514 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jgm8r" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.139112 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jgm8r"] Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.195523 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jct9r"] Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.195977 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jct9r" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="registry-server" containerID="cri-o://fd0cae8010626e20b9d43170aad30768d64b58d4a3e8e0398ac6ce023ab679da" gracePeriod=2 Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.383937 4906 generic.go:334] "Generic (PLEG): container finished" podID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerID="fd0cae8010626e20b9d43170aad30768d64b58d4a3e8e0398ac6ce023ab679da" exitCode=0 Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.384046 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jct9r" event={"ID":"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b","Type":"ContainerDied","Data":"fd0cae8010626e20b9d43170aad30768d64b58d4a3e8e0398ac6ce023ab679da"} Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.639598 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jct9r" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.775662 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv7rw\" (UniqueName: \"kubernetes.io/projected/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-kube-api-access-bv7rw\") pod \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.775772 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-utilities\") pod \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.775913 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-catalog-content\") pod \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\" (UID: \"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b\") " Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.776956 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-utilities" (OuterVolumeSpecName: "utilities") pod "5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" (UID: "5ab0bfa3-8e0b-4f95-a10b-57d2138f116b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.783183 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-kube-api-access-bv7rw" (OuterVolumeSpecName: "kube-api-access-bv7rw") pod "5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" (UID: "5ab0bfa3-8e0b-4f95-a10b-57d2138f116b"). InnerVolumeSpecName "kube-api-access-bv7rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.820458 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" (UID: "5ab0bfa3-8e0b-4f95-a10b-57d2138f116b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.877597 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.877638 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv7rw\" (UniqueName: \"kubernetes.io/projected/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-kube-api-access-bv7rw\") on node \"crc\" DevicePath \"\"" Nov 23 07:24:39 crc kubenswrapper[4906]: I1123 07:24:39.877653 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:24:40 crc kubenswrapper[4906]: I1123 07:24:40.395011 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jct9r" event={"ID":"5ab0bfa3-8e0b-4f95-a10b-57d2138f116b","Type":"ContainerDied","Data":"fac107a7b23cd07ea7151cff99ddabd7382710b7ef6bda53ee1ed99561ee6342"} Nov 23 07:24:40 crc kubenswrapper[4906]: I1123 07:24:40.395581 4906 scope.go:117] "RemoveContainer" containerID="fd0cae8010626e20b9d43170aad30768d64b58d4a3e8e0398ac6ce023ab679da" Nov 23 07:24:40 crc kubenswrapper[4906]: I1123 07:24:40.395271 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jct9r" Nov 23 07:24:40 crc kubenswrapper[4906]: I1123 07:24:40.427078 4906 scope.go:117] "RemoveContainer" containerID="735d9b427b1266406085d53a6e4b93cffeb42f5701f0bbdd9dd756d0b60ed523" Nov 23 07:24:40 crc kubenswrapper[4906]: I1123 07:24:40.432766 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jct9r"] Nov 23 07:24:40 crc kubenswrapper[4906]: I1123 07:24:40.437922 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jct9r"] Nov 23 07:24:40 crc kubenswrapper[4906]: I1123 07:24:40.469624 4906 scope.go:117] "RemoveContainer" containerID="5d3c532bd1b018440df527bdc4080286c82eee1497b9dff66a7a45f795ead44c" Nov 23 07:24:41 crc kubenswrapper[4906]: I1123 07:24:41.368309 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" path="/var/lib/kubelet/pods/5ab0bfa3-8e0b-4f95-a10b-57d2138f116b/volumes" Nov 23 07:25:20 crc kubenswrapper[4906]: I1123 07:25:20.945564 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:25:20 crc kubenswrapper[4906]: I1123 07:25:20.946353 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:25:50 crc kubenswrapper[4906]: I1123 07:25:50.945880 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:25:50 crc kubenswrapper[4906]: I1123 07:25:50.946630 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:26:20 crc kubenswrapper[4906]: I1123 07:26:20.946485 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:26:20 crc kubenswrapper[4906]: I1123 07:26:20.947613 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:26:20 crc kubenswrapper[4906]: I1123 07:26:20.947741 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:26:20 crc kubenswrapper[4906]: I1123 07:26:20.949167 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e38ec44b499cd9cd90308d44b4f055bb77948119e4077d14978ca4ea613b870"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:26:20 crc kubenswrapper[4906]: I1123 07:26:20.949281 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://7e38ec44b499cd9cd90308d44b4f055bb77948119e4077d14978ca4ea613b870" gracePeriod=600 Nov 23 07:26:21 crc kubenswrapper[4906]: I1123 07:26:21.584909 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="7e38ec44b499cd9cd90308d44b4f055bb77948119e4077d14978ca4ea613b870" exitCode=0 Nov 23 07:26:21 crc kubenswrapper[4906]: I1123 07:26:21.584957 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"7e38ec44b499cd9cd90308d44b4f055bb77948119e4077d14978ca4ea613b870"} Nov 23 07:26:21 crc kubenswrapper[4906]: I1123 07:26:21.585599 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a"} Nov 23 07:26:21 crc kubenswrapper[4906]: I1123 07:26:21.585653 4906 scope.go:117] "RemoveContainer" containerID="fdae235a54db512aba218f61d0f11bb013c4923669560566f0532db0e223f6a7" Nov 23 07:28:50 crc kubenswrapper[4906]: I1123 07:28:50.945783 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:28:50 crc kubenswrapper[4906]: I1123 07:28:50.946517 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.653696 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v4wj4"] Nov 23 07:29:14 crc kubenswrapper[4906]: E1123 07:29:14.656740 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="extract-content" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.656762 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="extract-content" Nov 23 07:29:14 crc kubenswrapper[4906]: E1123 07:29:14.656794 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="registry-server" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.656803 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="registry-server" Nov 23 07:29:14 crc kubenswrapper[4906]: E1123 07:29:14.656837 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="extract-utilities" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.656847 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="extract-utilities" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.657083 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab0bfa3-8e0b-4f95-a10b-57d2138f116b" containerName="registry-server" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.659534 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.691742 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4wj4"] Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.811744 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-utilities\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.811827 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhtmf\" (UniqueName: \"kubernetes.io/projected/a6e5971c-993e-4f5e-8f6e-fff9e514364a-kube-api-access-qhtmf\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.811870 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-catalog-content\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.913378 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-catalog-content\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.913518 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-utilities\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.913559 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhtmf\" (UniqueName: \"kubernetes.io/projected/a6e5971c-993e-4f5e-8f6e-fff9e514364a-kube-api-access-qhtmf\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.914090 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-catalog-content\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.914232 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-utilities\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:14 crc kubenswrapper[4906]: I1123 07:29:14.937553 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhtmf\" (UniqueName: \"kubernetes.io/projected/a6e5971c-993e-4f5e-8f6e-fff9e514364a-kube-api-access-qhtmf\") pod \"certified-operators-v4wj4\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:15 crc kubenswrapper[4906]: I1123 07:29:15.033666 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:15 crc kubenswrapper[4906]: I1123 07:29:15.510227 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4wj4"] Nov 23 07:29:16 crc kubenswrapper[4906]: I1123 07:29:16.014086 4906 generic.go:334] "Generic (PLEG): container finished" podID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerID="4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a" exitCode=0 Nov 23 07:29:16 crc kubenswrapper[4906]: I1123 07:29:16.014476 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4wj4" event={"ID":"a6e5971c-993e-4f5e-8f6e-fff9e514364a","Type":"ContainerDied","Data":"4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a"} Nov 23 07:29:16 crc kubenswrapper[4906]: I1123 07:29:16.014509 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4wj4" event={"ID":"a6e5971c-993e-4f5e-8f6e-fff9e514364a","Type":"ContainerStarted","Data":"c6e93befe71676e30c3ab4733c350678b26e3b0e4dfa93bfdfadbcf0a7847ac6"} Nov 23 07:29:16 crc kubenswrapper[4906]: I1123 07:29:16.017319 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:29:17 crc kubenswrapper[4906]: I1123 07:29:17.022693 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4wj4" event={"ID":"a6e5971c-993e-4f5e-8f6e-fff9e514364a","Type":"ContainerStarted","Data":"99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791"} Nov 23 07:29:18 crc kubenswrapper[4906]: I1123 07:29:18.037644 4906 generic.go:334] "Generic (PLEG): container finished" podID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerID="99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791" exitCode=0 Nov 23 07:29:18 crc kubenswrapper[4906]: I1123 07:29:18.038076 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4wj4" event={"ID":"a6e5971c-993e-4f5e-8f6e-fff9e514364a","Type":"ContainerDied","Data":"99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791"} Nov 23 07:29:19 crc kubenswrapper[4906]: I1123 07:29:19.052813 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4wj4" event={"ID":"a6e5971c-993e-4f5e-8f6e-fff9e514364a","Type":"ContainerStarted","Data":"7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2"} Nov 23 07:29:19 crc kubenswrapper[4906]: I1123 07:29:19.082211 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v4wj4" podStartSLOduration=2.612837378 podStartE2EDuration="5.082185671s" podCreationTimestamp="2025-11-23 07:29:14 +0000 UTC" firstStartedPulling="2025-11-23 07:29:16.017107921 +0000 UTC m=+2371.530499224" lastFinishedPulling="2025-11-23 07:29:18.486456204 +0000 UTC m=+2373.999847517" observedRunningTime="2025-11-23 07:29:19.080848936 +0000 UTC m=+2374.594240249" watchObservedRunningTime="2025-11-23 07:29:19.082185671 +0000 UTC m=+2374.595576984" Nov 23 07:29:20 crc kubenswrapper[4906]: I1123 07:29:20.946650 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:29:20 crc kubenswrapper[4906]: I1123 07:29:20.946818 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:29:25 crc kubenswrapper[4906]: I1123 07:29:25.034274 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:25 crc kubenswrapper[4906]: I1123 07:29:25.036246 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:25 crc kubenswrapper[4906]: I1123 07:29:25.099997 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:25 crc kubenswrapper[4906]: I1123 07:29:25.208232 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:25 crc kubenswrapper[4906]: I1123 07:29:25.341914 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4wj4"] Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.141953 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v4wj4" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="registry-server" containerID="cri-o://7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2" gracePeriod=2 Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.657445 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.788801 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-utilities\") pod \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.789071 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhtmf\" (UniqueName: \"kubernetes.io/projected/a6e5971c-993e-4f5e-8f6e-fff9e514364a-kube-api-access-qhtmf\") pod \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.789102 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-catalog-content\") pod \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\" (UID: \"a6e5971c-993e-4f5e-8f6e-fff9e514364a\") " Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.790762 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-utilities" (OuterVolumeSpecName: "utilities") pod "a6e5971c-993e-4f5e-8f6e-fff9e514364a" (UID: "a6e5971c-993e-4f5e-8f6e-fff9e514364a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.798247 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e5971c-993e-4f5e-8f6e-fff9e514364a-kube-api-access-qhtmf" (OuterVolumeSpecName: "kube-api-access-qhtmf") pod "a6e5971c-993e-4f5e-8f6e-fff9e514364a" (UID: "a6e5971c-993e-4f5e-8f6e-fff9e514364a"). InnerVolumeSpecName "kube-api-access-qhtmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.836752 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6e5971c-993e-4f5e-8f6e-fff9e514364a" (UID: "a6e5971c-993e-4f5e-8f6e-fff9e514364a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.892244 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhtmf\" (UniqueName: \"kubernetes.io/projected/a6e5971c-993e-4f5e-8f6e-fff9e514364a-kube-api-access-qhtmf\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.892336 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:27 crc kubenswrapper[4906]: I1123 07:29:27.892358 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e5971c-993e-4f5e-8f6e-fff9e514364a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.155448 4906 generic.go:334] "Generic (PLEG): container finished" podID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerID="7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2" exitCode=0 Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.155521 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4wj4" event={"ID":"a6e5971c-993e-4f5e-8f6e-fff9e514364a","Type":"ContainerDied","Data":"7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2"} Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.155584 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4wj4" event={"ID":"a6e5971c-993e-4f5e-8f6e-fff9e514364a","Type":"ContainerDied","Data":"c6e93befe71676e30c3ab4733c350678b26e3b0e4dfa93bfdfadbcf0a7847ac6"} Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.155589 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4wj4" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.155626 4906 scope.go:117] "RemoveContainer" containerID="7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.197866 4906 scope.go:117] "RemoveContainer" containerID="99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.238013 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4wj4"] Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.244955 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v4wj4"] Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.269112 4906 scope.go:117] "RemoveContainer" containerID="4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.301631 4906 scope.go:117] "RemoveContainer" containerID="7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2" Nov 23 07:29:28 crc kubenswrapper[4906]: E1123 07:29:28.302328 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2\": container with ID starting with 7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2 not found: ID does not exist" containerID="7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.302511 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2"} err="failed to get container status \"7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2\": rpc error: code = NotFound desc = could not find container \"7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2\": container with ID starting with 7a38b33071ff5bf511698c2e567b8ae5976109209f14bdd6ff5bcc4be9942bd2 not found: ID does not exist" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.302662 4906 scope.go:117] "RemoveContainer" containerID="99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791" Nov 23 07:29:28 crc kubenswrapper[4906]: E1123 07:29:28.303736 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791\": container with ID starting with 99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791 not found: ID does not exist" containerID="99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.303779 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791"} err="failed to get container status \"99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791\": rpc error: code = NotFound desc = could not find container \"99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791\": container with ID starting with 99697aba98171c0a1d3b798b4f08b035f0c32b0669564105a2cc900944b08791 not found: ID does not exist" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.303802 4906 scope.go:117] "RemoveContainer" containerID="4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a" Nov 23 07:29:28 crc kubenswrapper[4906]: E1123 07:29:28.304326 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a\": container with ID starting with 4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a not found: ID does not exist" containerID="4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a" Nov 23 07:29:28 crc kubenswrapper[4906]: I1123 07:29:28.304474 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a"} err="failed to get container status \"4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a\": rpc error: code = NotFound desc = could not find container \"4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a\": container with ID starting with 4764d59b1d4fdf5a44633fc30d61500036bf39d00a5520a46c740a58505f0b7a not found: ID does not exist" Nov 23 07:29:29 crc kubenswrapper[4906]: I1123 07:29:29.374083 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" path="/var/lib/kubelet/pods/a6e5971c-993e-4f5e-8f6e-fff9e514364a/volumes" Nov 23 07:29:50 crc kubenswrapper[4906]: I1123 07:29:50.946401 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:29:50 crc kubenswrapper[4906]: I1123 07:29:50.947189 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:29:50 crc kubenswrapper[4906]: I1123 07:29:50.947268 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:29:50 crc kubenswrapper[4906]: I1123 07:29:50.948378 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:29:50 crc kubenswrapper[4906]: I1123 07:29:50.948478 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" gracePeriod=600 Nov 23 07:29:51 crc kubenswrapper[4906]: E1123 07:29:51.083527 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:29:51 crc kubenswrapper[4906]: I1123 07:29:51.398500 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" exitCode=0 Nov 23 07:29:51 crc kubenswrapper[4906]: I1123 07:29:51.398560 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a"} Nov 23 07:29:51 crc kubenswrapper[4906]: I1123 07:29:51.398624 4906 scope.go:117] "RemoveContainer" containerID="7e38ec44b499cd9cd90308d44b4f055bb77948119e4077d14978ca4ea613b870" Nov 23 07:29:51 crc kubenswrapper[4906]: I1123 07:29:51.399540 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:29:51 crc kubenswrapper[4906]: E1123 07:29:51.400294 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.173956 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms"] Nov 23 07:30:00 crc kubenswrapper[4906]: E1123 07:30:00.175063 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="extract-utilities" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.175084 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="extract-utilities" Nov 23 07:30:00 crc kubenswrapper[4906]: E1123 07:30:00.175099 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="registry-server" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.175107 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="registry-server" Nov 23 07:30:00 crc kubenswrapper[4906]: E1123 07:30:00.175115 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="extract-content" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.175122 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="extract-content" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.175278 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6e5971c-993e-4f5e-8f6e-fff9e514364a" containerName="registry-server" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.176000 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.178936 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.189991 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms"] Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.191431 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.294852 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-secret-volume\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.294917 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p492s\" (UniqueName: \"kubernetes.io/projected/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-kube-api-access-p492s\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.295000 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-config-volume\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.396653 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-secret-volume\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.396725 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p492s\" (UniqueName: \"kubernetes.io/projected/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-kube-api-access-p492s\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.396787 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-config-volume\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.397749 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-config-volume\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.411720 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-secret-volume\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.411823 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p492s\" (UniqueName: \"kubernetes.io/projected/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-kube-api-access-p492s\") pod \"collect-profiles-29398050-ds2ms\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.508170 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:00 crc kubenswrapper[4906]: I1123 07:30:00.977066 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms"] Nov 23 07:30:01 crc kubenswrapper[4906]: I1123 07:30:01.498552 4906 generic.go:334] "Generic (PLEG): container finished" podID="80abab3e-830e-493f-8bb1-6bb9d8bd1a32" containerID="df20c20b9d37b1a5a5acf67ef09647ddf26fa72f65a94ae13f902a158708d4c1" exitCode=0 Nov 23 07:30:01 crc kubenswrapper[4906]: I1123 07:30:01.498640 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" event={"ID":"80abab3e-830e-493f-8bb1-6bb9d8bd1a32","Type":"ContainerDied","Data":"df20c20b9d37b1a5a5acf67ef09647ddf26fa72f65a94ae13f902a158708d4c1"} Nov 23 07:30:01 crc kubenswrapper[4906]: I1123 07:30:01.498995 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" event={"ID":"80abab3e-830e-493f-8bb1-6bb9d8bd1a32","Type":"ContainerStarted","Data":"d38e7a3c21540b218f6ec58307d44a85c99cff3d58e694a7f8dbcd220a80cf2e"} Nov 23 07:30:02 crc kubenswrapper[4906]: I1123 07:30:02.826868 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:02 crc kubenswrapper[4906]: I1123 07:30:02.963364 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-secret-volume\") pod \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " Nov 23 07:30:02 crc kubenswrapper[4906]: I1123 07:30:02.963455 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-config-volume\") pod \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " Nov 23 07:30:02 crc kubenswrapper[4906]: I1123 07:30:02.963544 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p492s\" (UniqueName: \"kubernetes.io/projected/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-kube-api-access-p492s\") pod \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\" (UID: \"80abab3e-830e-493f-8bb1-6bb9d8bd1a32\") " Nov 23 07:30:02 crc kubenswrapper[4906]: I1123 07:30:02.966589 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-config-volume" (OuterVolumeSpecName: "config-volume") pod "80abab3e-830e-493f-8bb1-6bb9d8bd1a32" (UID: "80abab3e-830e-493f-8bb1-6bb9d8bd1a32"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:30:02 crc kubenswrapper[4906]: I1123 07:30:02.971767 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-kube-api-access-p492s" (OuterVolumeSpecName: "kube-api-access-p492s") pod "80abab3e-830e-493f-8bb1-6bb9d8bd1a32" (UID: "80abab3e-830e-493f-8bb1-6bb9d8bd1a32"). InnerVolumeSpecName "kube-api-access-p492s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:30:02 crc kubenswrapper[4906]: I1123 07:30:02.972380 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "80abab3e-830e-493f-8bb1-6bb9d8bd1a32" (UID: "80abab3e-830e-493f-8bb1-6bb9d8bd1a32"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.065931 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.065972 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.065983 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p492s\" (UniqueName: \"kubernetes.io/projected/80abab3e-830e-493f-8bb1-6bb9d8bd1a32-kube-api-access-p492s\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.518286 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" event={"ID":"80abab3e-830e-493f-8bb1-6bb9d8bd1a32","Type":"ContainerDied","Data":"d38e7a3c21540b218f6ec58307d44a85c99cff3d58e694a7f8dbcd220a80cf2e"} Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.518357 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d38e7a3c21540b218f6ec58307d44a85c99cff3d58e694a7f8dbcd220a80cf2e" Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.519026 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms" Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.930976 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx"] Nov 23 07:30:03 crc kubenswrapper[4906]: I1123 07:30:03.942569 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-pzgzx"] Nov 23 07:30:05 crc kubenswrapper[4906]: I1123 07:30:05.363016 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:30:05 crc kubenswrapper[4906]: E1123 07:30:05.363723 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:30:05 crc kubenswrapper[4906]: I1123 07:30:05.377438 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4" path="/var/lib/kubelet/pods/d92b911b-cee3-4ccd-8a2c-76dbe0d54fb4/volumes" Nov 23 07:30:20 crc kubenswrapper[4906]: I1123 07:30:20.357977 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:30:20 crc kubenswrapper[4906]: E1123 07:30:20.359184 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:30:33 crc kubenswrapper[4906]: I1123 07:30:33.357605 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:30:33 crc kubenswrapper[4906]: E1123 07:30:33.358871 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:30:48 crc kubenswrapper[4906]: I1123 07:30:48.356788 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:30:48 crc kubenswrapper[4906]: E1123 07:30:48.358038 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:30:56 crc kubenswrapper[4906]: I1123 07:30:56.749888 4906 scope.go:117] "RemoveContainer" containerID="e5b2a5b84a61e4bfd832cb93b9ffcba1179e59a07dcc491ee2c0d19826e3d971" Nov 23 07:31:01 crc kubenswrapper[4906]: I1123 07:31:01.357422 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:31:01 crc kubenswrapper[4906]: E1123 07:31:01.359046 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:31:14 crc kubenswrapper[4906]: I1123 07:31:14.356852 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:31:14 crc kubenswrapper[4906]: E1123 07:31:14.357619 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:31:28 crc kubenswrapper[4906]: I1123 07:31:28.356554 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:31:28 crc kubenswrapper[4906]: E1123 07:31:28.357901 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:31:39 crc kubenswrapper[4906]: I1123 07:31:39.358352 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:31:39 crc kubenswrapper[4906]: E1123 07:31:39.359704 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:31:54 crc kubenswrapper[4906]: I1123 07:31:54.356657 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:31:54 crc kubenswrapper[4906]: E1123 07:31:54.359432 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:32:08 crc kubenswrapper[4906]: I1123 07:32:08.357196 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:32:08 crc kubenswrapper[4906]: E1123 07:32:08.358300 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:32:22 crc kubenswrapper[4906]: I1123 07:32:22.358900 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:32:22 crc kubenswrapper[4906]: E1123 07:32:22.360142 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:32:34 crc kubenswrapper[4906]: I1123 07:32:34.356662 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:32:34 crc kubenswrapper[4906]: E1123 07:32:34.357831 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:32:45 crc kubenswrapper[4906]: I1123 07:32:45.362471 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:32:45 crc kubenswrapper[4906]: E1123 07:32:45.365181 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:33:00 crc kubenswrapper[4906]: I1123 07:33:00.357387 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:33:00 crc kubenswrapper[4906]: E1123 07:33:00.358926 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:33:11 crc kubenswrapper[4906]: I1123 07:33:11.357035 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:33:11 crc kubenswrapper[4906]: E1123 07:33:11.357873 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.184185 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lbd8n"] Nov 23 07:33:25 crc kubenswrapper[4906]: E1123 07:33:25.185423 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80abab3e-830e-493f-8bb1-6bb9d8bd1a32" containerName="collect-profiles" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.185449 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="80abab3e-830e-493f-8bb1-6bb9d8bd1a32" containerName="collect-profiles" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.185762 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="80abab3e-830e-493f-8bb1-6bb9d8bd1a32" containerName="collect-profiles" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.192012 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.204092 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbd8n"] Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.273401 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-catalog-content\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.273546 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64pk4\" (UniqueName: \"kubernetes.io/projected/3b0fca62-212f-414e-ab5b-29c04fedf3b4-kube-api-access-64pk4\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.273784 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-utilities\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.366226 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:33:25 crc kubenswrapper[4906]: E1123 07:33:25.366519 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.375594 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64pk4\" (UniqueName: \"kubernetes.io/projected/3b0fca62-212f-414e-ab5b-29c04fedf3b4-kube-api-access-64pk4\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.375733 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-utilities\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.375807 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-catalog-content\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.376388 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-utilities\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.376471 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-catalog-content\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.408420 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64pk4\" (UniqueName: \"kubernetes.io/projected/3b0fca62-212f-414e-ab5b-29c04fedf3b4-kube-api-access-64pk4\") pod \"redhat-marketplace-lbd8n\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:25 crc kubenswrapper[4906]: I1123 07:33:25.532264 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:26 crc kubenswrapper[4906]: I1123 07:33:26.006725 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbd8n"] Nov 23 07:33:26 crc kubenswrapper[4906]: I1123 07:33:26.538262 4906 generic.go:334] "Generic (PLEG): container finished" podID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerID="7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c" exitCode=0 Nov 23 07:33:26 crc kubenswrapper[4906]: I1123 07:33:26.538337 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbd8n" event={"ID":"3b0fca62-212f-414e-ab5b-29c04fedf3b4","Type":"ContainerDied","Data":"7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c"} Nov 23 07:33:26 crc kubenswrapper[4906]: I1123 07:33:26.538379 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbd8n" event={"ID":"3b0fca62-212f-414e-ab5b-29c04fedf3b4","Type":"ContainerStarted","Data":"e55ad448ba37944c72803a3f0142a5070cf078bccacb7a9663c1a52be463d628"} Nov 23 07:33:27 crc kubenswrapper[4906]: I1123 07:33:27.549155 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbd8n" event={"ID":"3b0fca62-212f-414e-ab5b-29c04fedf3b4","Type":"ContainerStarted","Data":"321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c"} Nov 23 07:33:28 crc kubenswrapper[4906]: I1123 07:33:28.559901 4906 generic.go:334] "Generic (PLEG): container finished" podID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerID="321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c" exitCode=0 Nov 23 07:33:28 crc kubenswrapper[4906]: I1123 07:33:28.559966 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbd8n" event={"ID":"3b0fca62-212f-414e-ab5b-29c04fedf3b4","Type":"ContainerDied","Data":"321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c"} Nov 23 07:33:29 crc kubenswrapper[4906]: I1123 07:33:29.574799 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbd8n" event={"ID":"3b0fca62-212f-414e-ab5b-29c04fedf3b4","Type":"ContainerStarted","Data":"7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88"} Nov 23 07:33:29 crc kubenswrapper[4906]: I1123 07:33:29.608000 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lbd8n" podStartSLOduration=2.224528693 podStartE2EDuration="4.607962528s" podCreationTimestamp="2025-11-23 07:33:25 +0000 UTC" firstStartedPulling="2025-11-23 07:33:26.544190579 +0000 UTC m=+2622.057581892" lastFinishedPulling="2025-11-23 07:33:28.927624414 +0000 UTC m=+2624.441015727" observedRunningTime="2025-11-23 07:33:29.606128439 +0000 UTC m=+2625.119519752" watchObservedRunningTime="2025-11-23 07:33:29.607962528 +0000 UTC m=+2625.121353871" Nov 23 07:33:35 crc kubenswrapper[4906]: I1123 07:33:35.532838 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:35 crc kubenswrapper[4906]: I1123 07:33:35.534094 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:35 crc kubenswrapper[4906]: I1123 07:33:35.607893 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:35 crc kubenswrapper[4906]: I1123 07:33:35.689913 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:35 crc kubenswrapper[4906]: I1123 07:33:35.860678 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbd8n"] Nov 23 07:33:36 crc kubenswrapper[4906]: I1123 07:33:36.357427 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:33:36 crc kubenswrapper[4906]: E1123 07:33:36.357754 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:33:37 crc kubenswrapper[4906]: I1123 07:33:37.643482 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lbd8n" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="registry-server" containerID="cri-o://7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88" gracePeriod=2 Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.071611 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.221772 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64pk4\" (UniqueName: \"kubernetes.io/projected/3b0fca62-212f-414e-ab5b-29c04fedf3b4-kube-api-access-64pk4\") pod \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.221896 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-utilities\") pod \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.222080 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-catalog-content\") pod \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\" (UID: \"3b0fca62-212f-414e-ab5b-29c04fedf3b4\") " Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.228324 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0fca62-212f-414e-ab5b-29c04fedf3b4-kube-api-access-64pk4" (OuterVolumeSpecName: "kube-api-access-64pk4") pod "3b0fca62-212f-414e-ab5b-29c04fedf3b4" (UID: "3b0fca62-212f-414e-ab5b-29c04fedf3b4"). InnerVolumeSpecName "kube-api-access-64pk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.228712 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-utilities" (OuterVolumeSpecName: "utilities") pod "3b0fca62-212f-414e-ab5b-29c04fedf3b4" (UID: "3b0fca62-212f-414e-ab5b-29c04fedf3b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.252864 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b0fca62-212f-414e-ab5b-29c04fedf3b4" (UID: "3b0fca62-212f-414e-ab5b-29c04fedf3b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.324419 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.324465 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64pk4\" (UniqueName: \"kubernetes.io/projected/3b0fca62-212f-414e-ab5b-29c04fedf3b4-kube-api-access-64pk4\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.324484 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b0fca62-212f-414e-ab5b-29c04fedf3b4-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.658792 4906 generic.go:334] "Generic (PLEG): container finished" podID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerID="7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88" exitCode=0 Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.658942 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbd8n" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.658928 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbd8n" event={"ID":"3b0fca62-212f-414e-ab5b-29c04fedf3b4","Type":"ContainerDied","Data":"7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88"} Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.659048 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbd8n" event={"ID":"3b0fca62-212f-414e-ab5b-29c04fedf3b4","Type":"ContainerDied","Data":"e55ad448ba37944c72803a3f0142a5070cf078bccacb7a9663c1a52be463d628"} Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.659115 4906 scope.go:117] "RemoveContainer" containerID="7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.691476 4906 scope.go:117] "RemoveContainer" containerID="321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.718341 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbd8n"] Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.725049 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbd8n"] Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.738535 4906 scope.go:117] "RemoveContainer" containerID="7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.767539 4906 scope.go:117] "RemoveContainer" containerID="7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88" Nov 23 07:33:38 crc kubenswrapper[4906]: E1123 07:33:38.768523 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88\": container with ID starting with 7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88 not found: ID does not exist" containerID="7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.768612 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88"} err="failed to get container status \"7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88\": rpc error: code = NotFound desc = could not find container \"7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88\": container with ID starting with 7218bc7c38bec17361306513d6cd002e6e877d5195fa12c609abe10defc97a88 not found: ID does not exist" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.768666 4906 scope.go:117] "RemoveContainer" containerID="321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c" Nov 23 07:33:38 crc kubenswrapper[4906]: E1123 07:33:38.769378 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c\": container with ID starting with 321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c not found: ID does not exist" containerID="321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.769421 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c"} err="failed to get container status \"321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c\": rpc error: code = NotFound desc = could not find container \"321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c\": container with ID starting with 321bf115e962154f13b341bbbbcd244b7b0350bd4cc85487b973f42579d7d86c not found: ID does not exist" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.769456 4906 scope.go:117] "RemoveContainer" containerID="7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c" Nov 23 07:33:38 crc kubenswrapper[4906]: E1123 07:33:38.770002 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c\": container with ID starting with 7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c not found: ID does not exist" containerID="7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c" Nov 23 07:33:38 crc kubenswrapper[4906]: I1123 07:33:38.770050 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c"} err="failed to get container status \"7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c\": rpc error: code = NotFound desc = could not find container \"7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c\": container with ID starting with 7f72c5b743b15b3353509c542f81b98744e20b1a594a6aba36611795fb4e965c not found: ID does not exist" Nov 23 07:33:39 crc kubenswrapper[4906]: I1123 07:33:39.373788 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" path="/var/lib/kubelet/pods/3b0fca62-212f-414e-ab5b-29c04fedf3b4/volumes" Nov 23 07:33:47 crc kubenswrapper[4906]: I1123 07:33:47.356643 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:33:47 crc kubenswrapper[4906]: E1123 07:33:47.357639 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:33:58 crc kubenswrapper[4906]: I1123 07:33:58.356908 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:33:58 crc kubenswrapper[4906]: E1123 07:33:58.358034 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:34:11 crc kubenswrapper[4906]: I1123 07:34:11.356845 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:34:11 crc kubenswrapper[4906]: E1123 07:34:11.357794 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.261748 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tdskf"] Nov 23 07:34:22 crc kubenswrapper[4906]: E1123 07:34:22.263017 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="extract-content" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.263037 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="extract-content" Nov 23 07:34:22 crc kubenswrapper[4906]: E1123 07:34:22.263058 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="registry-server" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.263066 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="registry-server" Nov 23 07:34:22 crc kubenswrapper[4906]: E1123 07:34:22.263083 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="extract-utilities" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.263093 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="extract-utilities" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.263286 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0fca62-212f-414e-ab5b-29c04fedf3b4" containerName="registry-server" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.265084 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.283358 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdskf"] Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.345424 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgp68\" (UniqueName: \"kubernetes.io/projected/2951c6c5-defe-475a-a09b-3ef4e672e420-kube-api-access-vgp68\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.345514 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-catalog-content\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.345569 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-utilities\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.446927 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgp68\" (UniqueName: \"kubernetes.io/projected/2951c6c5-defe-475a-a09b-3ef4e672e420-kube-api-access-vgp68\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.447016 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-catalog-content\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.447065 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-utilities\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.447549 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-utilities\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.447851 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-catalog-content\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.489218 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgp68\" (UniqueName: \"kubernetes.io/projected/2951c6c5-defe-475a-a09b-3ef4e672e420-kube-api-access-vgp68\") pod \"community-operators-tdskf\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:22 crc kubenswrapper[4906]: I1123 07:34:22.637469 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:23 crc kubenswrapper[4906]: I1123 07:34:23.145036 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdskf"] Nov 23 07:34:24 crc kubenswrapper[4906]: I1123 07:34:24.131059 4906 generic.go:334] "Generic (PLEG): container finished" podID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerID="2ba31e25749f31a8bbf77f3d92c26e540d8ce2eb5914fc10f98976a38a61c8d7" exitCode=0 Nov 23 07:34:24 crc kubenswrapper[4906]: I1123 07:34:24.131133 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdskf" event={"ID":"2951c6c5-defe-475a-a09b-3ef4e672e420","Type":"ContainerDied","Data":"2ba31e25749f31a8bbf77f3d92c26e540d8ce2eb5914fc10f98976a38a61c8d7"} Nov 23 07:34:24 crc kubenswrapper[4906]: I1123 07:34:24.131182 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdskf" event={"ID":"2951c6c5-defe-475a-a09b-3ef4e672e420","Type":"ContainerStarted","Data":"91e6c13996300674b6ee3b1ff6277ab2f9b99b80cf7dc39545b16373f981dd38"} Nov 23 07:34:24 crc kubenswrapper[4906]: I1123 07:34:24.135940 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:34:25 crc kubenswrapper[4906]: I1123 07:34:25.147252 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdskf" event={"ID":"2951c6c5-defe-475a-a09b-3ef4e672e420","Type":"ContainerStarted","Data":"64cdf1bf11eac261b8c69ac4cd3b7f29e248fd362db7a6f11816e6093166385d"} Nov 23 07:34:25 crc kubenswrapper[4906]: I1123 07:34:25.363530 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:34:25 crc kubenswrapper[4906]: E1123 07:34:25.364088 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:34:26 crc kubenswrapper[4906]: I1123 07:34:26.160535 4906 generic.go:334] "Generic (PLEG): container finished" podID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerID="64cdf1bf11eac261b8c69ac4cd3b7f29e248fd362db7a6f11816e6093166385d" exitCode=0 Nov 23 07:34:26 crc kubenswrapper[4906]: I1123 07:34:26.160644 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdskf" event={"ID":"2951c6c5-defe-475a-a09b-3ef4e672e420","Type":"ContainerDied","Data":"64cdf1bf11eac261b8c69ac4cd3b7f29e248fd362db7a6f11816e6093166385d"} Nov 23 07:34:27 crc kubenswrapper[4906]: I1123 07:34:27.171631 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdskf" event={"ID":"2951c6c5-defe-475a-a09b-3ef4e672e420","Type":"ContainerStarted","Data":"f743b0d4e383bb5e8166a426fc8cbd0924746f8a4e2f1e02d2eeca90313d734c"} Nov 23 07:34:27 crc kubenswrapper[4906]: I1123 07:34:27.200572 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tdskf" podStartSLOduration=2.640368443 podStartE2EDuration="5.200547349s" podCreationTimestamp="2025-11-23 07:34:22 +0000 UTC" firstStartedPulling="2025-11-23 07:34:24.13524289 +0000 UTC m=+2679.648634233" lastFinishedPulling="2025-11-23 07:34:26.695421786 +0000 UTC m=+2682.208813139" observedRunningTime="2025-11-23 07:34:27.198217228 +0000 UTC m=+2682.711608571" watchObservedRunningTime="2025-11-23 07:34:27.200547349 +0000 UTC m=+2682.713938652" Nov 23 07:34:32 crc kubenswrapper[4906]: I1123 07:34:32.637647 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:32 crc kubenswrapper[4906]: I1123 07:34:32.638557 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:32 crc kubenswrapper[4906]: I1123 07:34:32.772760 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.287586 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.452947 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r99w4"] Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.455610 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.483901 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r99w4"] Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.554512 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cwgn\" (UniqueName: \"kubernetes.io/projected/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-kube-api-access-4cwgn\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.554660 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-utilities\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.554795 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-catalog-content\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.656377 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cwgn\" (UniqueName: \"kubernetes.io/projected/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-kube-api-access-4cwgn\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.656761 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-utilities\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.656844 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-catalog-content\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.657939 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-catalog-content\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.659033 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-utilities\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.687556 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cwgn\" (UniqueName: \"kubernetes.io/projected/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-kube-api-access-4cwgn\") pod \"redhat-operators-r99w4\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:33 crc kubenswrapper[4906]: I1123 07:34:33.805773 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:34 crc kubenswrapper[4906]: I1123 07:34:34.088614 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r99w4"] Nov 23 07:34:34 crc kubenswrapper[4906]: I1123 07:34:34.244607 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99w4" event={"ID":"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60","Type":"ContainerStarted","Data":"f171d6c8739e0e61879777d4b5f85af2bcb64fdf40e07160c086bbfb9faeeefe"} Nov 23 07:34:35 crc kubenswrapper[4906]: I1123 07:34:35.262986 4906 generic.go:334] "Generic (PLEG): container finished" podID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerID="ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d" exitCode=0 Nov 23 07:34:35 crc kubenswrapper[4906]: I1123 07:34:35.263058 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99w4" event={"ID":"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60","Type":"ContainerDied","Data":"ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d"} Nov 23 07:34:35 crc kubenswrapper[4906]: I1123 07:34:35.841177 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdskf"] Nov 23 07:34:35 crc kubenswrapper[4906]: I1123 07:34:35.842139 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tdskf" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="registry-server" containerID="cri-o://f743b0d4e383bb5e8166a426fc8cbd0924746f8a4e2f1e02d2eeca90313d734c" gracePeriod=2 Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.279237 4906 generic.go:334] "Generic (PLEG): container finished" podID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerID="f743b0d4e383bb5e8166a426fc8cbd0924746f8a4e2f1e02d2eeca90313d734c" exitCode=0 Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.279376 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdskf" event={"ID":"2951c6c5-defe-475a-a09b-3ef4e672e420","Type":"ContainerDied","Data":"f743b0d4e383bb5e8166a426fc8cbd0924746f8a4e2f1e02d2eeca90313d734c"} Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.282320 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99w4" event={"ID":"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60","Type":"ContainerStarted","Data":"7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296"} Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.330900 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.358489 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:34:36 crc kubenswrapper[4906]: E1123 07:34:36.358941 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.406078 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-catalog-content\") pod \"2951c6c5-defe-475a-a09b-3ef4e672e420\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.406265 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgp68\" (UniqueName: \"kubernetes.io/projected/2951c6c5-defe-475a-a09b-3ef4e672e420-kube-api-access-vgp68\") pod \"2951c6c5-defe-475a-a09b-3ef4e672e420\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.424978 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2951c6c5-defe-475a-a09b-3ef4e672e420-kube-api-access-vgp68" (OuterVolumeSpecName: "kube-api-access-vgp68") pod "2951c6c5-defe-475a-a09b-3ef4e672e420" (UID: "2951c6c5-defe-475a-a09b-3ef4e672e420"). InnerVolumeSpecName "kube-api-access-vgp68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.496263 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2951c6c5-defe-475a-a09b-3ef4e672e420" (UID: "2951c6c5-defe-475a-a09b-3ef4e672e420"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.508249 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-utilities\") pod \"2951c6c5-defe-475a-a09b-3ef4e672e420\" (UID: \"2951c6c5-defe-475a-a09b-3ef4e672e420\") " Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.508818 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.508839 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgp68\" (UniqueName: \"kubernetes.io/projected/2951c6c5-defe-475a-a09b-3ef4e672e420-kube-api-access-vgp68\") on node \"crc\" DevicePath \"\"" Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.509845 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-utilities" (OuterVolumeSpecName: "utilities") pod "2951c6c5-defe-475a-a09b-3ef4e672e420" (UID: "2951c6c5-defe-475a-a09b-3ef4e672e420"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:34:36 crc kubenswrapper[4906]: I1123 07:34:36.610301 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2951c6c5-defe-475a-a09b-3ef4e672e420-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.296273 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdskf" Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.296275 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdskf" event={"ID":"2951c6c5-defe-475a-a09b-3ef4e672e420","Type":"ContainerDied","Data":"91e6c13996300674b6ee3b1ff6277ab2f9b99b80cf7dc39545b16373f981dd38"} Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.296377 4906 scope.go:117] "RemoveContainer" containerID="f743b0d4e383bb5e8166a426fc8cbd0924746f8a4e2f1e02d2eeca90313d734c" Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.298981 4906 generic.go:334] "Generic (PLEG): container finished" podID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerID="7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296" exitCode=0 Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.299028 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99w4" event={"ID":"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60","Type":"ContainerDied","Data":"7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296"} Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.324462 4906 scope.go:117] "RemoveContainer" containerID="64cdf1bf11eac261b8c69ac4cd3b7f29e248fd362db7a6f11816e6093166385d" Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.372319 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdskf"] Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.372406 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tdskf"] Nov 23 07:34:37 crc kubenswrapper[4906]: I1123 07:34:37.391191 4906 scope.go:117] "RemoveContainer" containerID="2ba31e25749f31a8bbf77f3d92c26e540d8ce2eb5914fc10f98976a38a61c8d7" Nov 23 07:34:38 crc kubenswrapper[4906]: I1123 07:34:38.311879 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99w4" event={"ID":"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60","Type":"ContainerStarted","Data":"d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0"} Nov 23 07:34:38 crc kubenswrapper[4906]: I1123 07:34:38.344076 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r99w4" podStartSLOduration=2.933127584 podStartE2EDuration="5.344050093s" podCreationTimestamp="2025-11-23 07:34:33 +0000 UTC" firstStartedPulling="2025-11-23 07:34:35.265926625 +0000 UTC m=+2690.779317968" lastFinishedPulling="2025-11-23 07:34:37.676849134 +0000 UTC m=+2693.190240477" observedRunningTime="2025-11-23 07:34:38.334839079 +0000 UTC m=+2693.848230412" watchObservedRunningTime="2025-11-23 07:34:38.344050093 +0000 UTC m=+2693.857441436" Nov 23 07:34:39 crc kubenswrapper[4906]: I1123 07:34:39.373709 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" path="/var/lib/kubelet/pods/2951c6c5-defe-475a-a09b-3ef4e672e420/volumes" Nov 23 07:34:43 crc kubenswrapper[4906]: I1123 07:34:43.806467 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:43 crc kubenswrapper[4906]: I1123 07:34:43.807375 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:44 crc kubenswrapper[4906]: I1123 07:34:44.869888 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r99w4" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="registry-server" probeResult="failure" output=< Nov 23 07:34:44 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 07:34:44 crc kubenswrapper[4906]: > Nov 23 07:34:50 crc kubenswrapper[4906]: I1123 07:34:50.358157 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:34:50 crc kubenswrapper[4906]: E1123 07:34:50.359510 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:34:53 crc kubenswrapper[4906]: I1123 07:34:53.875856 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:53 crc kubenswrapper[4906]: I1123 07:34:53.934893 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:54 crc kubenswrapper[4906]: I1123 07:34:54.114503 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r99w4"] Nov 23 07:34:55 crc kubenswrapper[4906]: I1123 07:34:55.508779 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r99w4" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="registry-server" containerID="cri-o://d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0" gracePeriod=2 Nov 23 07:34:55 crc kubenswrapper[4906]: I1123 07:34:55.926405 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.099946 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-catalog-content\") pod \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.100046 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-utilities\") pod \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.100119 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cwgn\" (UniqueName: \"kubernetes.io/projected/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-kube-api-access-4cwgn\") pod \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\" (UID: \"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60\") " Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.102430 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-utilities" (OuterVolumeSpecName: "utilities") pod "633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" (UID: "633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.110306 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-kube-api-access-4cwgn" (OuterVolumeSpecName: "kube-api-access-4cwgn") pod "633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" (UID: "633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60"). InnerVolumeSpecName "kube-api-access-4cwgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.202998 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.203057 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cwgn\" (UniqueName: \"kubernetes.io/projected/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-kube-api-access-4cwgn\") on node \"crc\" DevicePath \"\"" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.233171 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" (UID: "633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.305218 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.518731 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99w4" event={"ID":"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60","Type":"ContainerDied","Data":"d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0"} Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.518773 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r99w4" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.518826 4906 scope.go:117] "RemoveContainer" containerID="d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.518596 4906 generic.go:334] "Generic (PLEG): container finished" podID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerID="d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0" exitCode=0 Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.520769 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r99w4" event={"ID":"633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60","Type":"ContainerDied","Data":"f171d6c8739e0e61879777d4b5f85af2bcb64fdf40e07160c086bbfb9faeeefe"} Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.535250 4906 scope.go:117] "RemoveContainer" containerID="7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.559438 4906 scope.go:117] "RemoveContainer" containerID="ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.562511 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r99w4"] Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.568941 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r99w4"] Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.581220 4906 scope.go:117] "RemoveContainer" containerID="d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0" Nov 23 07:34:56 crc kubenswrapper[4906]: E1123 07:34:56.581738 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0\": container with ID starting with d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0 not found: ID does not exist" containerID="d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.581773 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0"} err="failed to get container status \"d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0\": rpc error: code = NotFound desc = could not find container \"d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0\": container with ID starting with d686ac1712370309a2b97d7727fd4bafc9ec322df5eab6dafe819b08b45167f0 not found: ID does not exist" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.581801 4906 scope.go:117] "RemoveContainer" containerID="7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296" Nov 23 07:34:56 crc kubenswrapper[4906]: E1123 07:34:56.582139 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296\": container with ID starting with 7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296 not found: ID does not exist" containerID="7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.582175 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296"} err="failed to get container status \"7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296\": rpc error: code = NotFound desc = could not find container \"7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296\": container with ID starting with 7be7590c3a8c1b80222d63ab5eb8d43e60c2abbbc8a2b32d718fe04ce0b05296 not found: ID does not exist" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.582195 4906 scope.go:117] "RemoveContainer" containerID="ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d" Nov 23 07:34:56 crc kubenswrapper[4906]: E1123 07:34:56.582743 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d\": container with ID starting with ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d not found: ID does not exist" containerID="ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d" Nov 23 07:34:56 crc kubenswrapper[4906]: I1123 07:34:56.582770 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d"} err="failed to get container status \"ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d\": rpc error: code = NotFound desc = could not find container \"ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d\": container with ID starting with ff6d24792c4764fae7cbc85b86965f1d97f7cd03e86b49c320dc0e45c5d8600d not found: ID does not exist" Nov 23 07:34:57 crc kubenswrapper[4906]: I1123 07:34:57.365663 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" path="/var/lib/kubelet/pods/633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60/volumes" Nov 23 07:35:03 crc kubenswrapper[4906]: I1123 07:35:03.356969 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:35:03 crc kubenswrapper[4906]: I1123 07:35:03.578735 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"20f1a767b5b092456875c9989d4040177d1080271980f988e48b049e61bf2e20"} Nov 23 07:37:20 crc kubenswrapper[4906]: I1123 07:37:20.946846 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:37:20 crc kubenswrapper[4906]: I1123 07:37:20.948069 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:37:50 crc kubenswrapper[4906]: I1123 07:37:50.946809 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:37:50 crc kubenswrapper[4906]: I1123 07:37:50.947510 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:38:20 crc kubenswrapper[4906]: I1123 07:38:20.946180 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:38:20 crc kubenswrapper[4906]: I1123 07:38:20.947063 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:38:20 crc kubenswrapper[4906]: I1123 07:38:20.947134 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:38:20 crc kubenswrapper[4906]: I1123 07:38:20.947945 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20f1a767b5b092456875c9989d4040177d1080271980f988e48b049e61bf2e20"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:38:20 crc kubenswrapper[4906]: I1123 07:38:20.948049 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://20f1a767b5b092456875c9989d4040177d1080271980f988e48b049e61bf2e20" gracePeriod=600 Nov 23 07:38:21 crc kubenswrapper[4906]: I1123 07:38:21.575930 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="20f1a767b5b092456875c9989d4040177d1080271980f988e48b049e61bf2e20" exitCode=0 Nov 23 07:38:21 crc kubenswrapper[4906]: I1123 07:38:21.576005 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"20f1a767b5b092456875c9989d4040177d1080271980f988e48b049e61bf2e20"} Nov 23 07:38:21 crc kubenswrapper[4906]: I1123 07:38:21.576230 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2"} Nov 23 07:38:21 crc kubenswrapper[4906]: I1123 07:38:21.576255 4906 scope.go:117] "RemoveContainer" containerID="89fb88470ff765cfaae7b4eb38c39340c8f0a5468c56d1676f0dee46eb1a5c0a" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.507281 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d7sxb"] Nov 23 07:40:41 crc kubenswrapper[4906]: E1123 07:40:41.508816 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="extract-content" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.508842 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="extract-content" Nov 23 07:40:41 crc kubenswrapper[4906]: E1123 07:40:41.508904 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="registry-server" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.508923 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="registry-server" Nov 23 07:40:41 crc kubenswrapper[4906]: E1123 07:40:41.508951 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="registry-server" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.508965 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="registry-server" Nov 23 07:40:41 crc kubenswrapper[4906]: E1123 07:40:41.508998 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="extract-content" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.509010 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="extract-content" Nov 23 07:40:41 crc kubenswrapper[4906]: E1123 07:40:41.509028 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="extract-utilities" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.509042 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="extract-utilities" Nov 23 07:40:41 crc kubenswrapper[4906]: E1123 07:40:41.509066 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="extract-utilities" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.509078 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="extract-utilities" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.509356 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2951c6c5-defe-475a-a09b-3ef4e672e420" containerName="registry-server" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.509404 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="633dd4c7-b38f-4e2b-9a1e-c6c61f8bab60" containerName="registry-server" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.511417 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.521190 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d7sxb"] Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.605949 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-catalog-content\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.606041 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjgv8\" (UniqueName: \"kubernetes.io/projected/bb39505c-0ba6-465c-a015-d5be03a5b03e-kube-api-access-qjgv8\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.606103 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-utilities\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.707426 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjgv8\" (UniqueName: \"kubernetes.io/projected/bb39505c-0ba6-465c-a015-d5be03a5b03e-kube-api-access-qjgv8\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.707508 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-utilities\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.707572 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-catalog-content\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.708172 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-catalog-content\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.708283 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-utilities\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.747028 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjgv8\" (UniqueName: \"kubernetes.io/projected/bb39505c-0ba6-465c-a015-d5be03a5b03e-kube-api-access-qjgv8\") pod \"certified-operators-d7sxb\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:41 crc kubenswrapper[4906]: I1123 07:40:41.847709 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:42 crc kubenswrapper[4906]: I1123 07:40:42.325819 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d7sxb"] Nov 23 07:40:43 crc kubenswrapper[4906]: I1123 07:40:43.088466 4906 generic.go:334] "Generic (PLEG): container finished" podID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerID="9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0" exitCode=0 Nov 23 07:40:43 crc kubenswrapper[4906]: I1123 07:40:43.088511 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7sxb" event={"ID":"bb39505c-0ba6-465c-a015-d5be03a5b03e","Type":"ContainerDied","Data":"9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0"} Nov 23 07:40:43 crc kubenswrapper[4906]: I1123 07:40:43.088540 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7sxb" event={"ID":"bb39505c-0ba6-465c-a015-d5be03a5b03e","Type":"ContainerStarted","Data":"a46213a39ef45e026ea3508bcd808b4fb1840e7ff40698ff44585f8adddc8f23"} Nov 23 07:40:43 crc kubenswrapper[4906]: I1123 07:40:43.091437 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:40:44 crc kubenswrapper[4906]: I1123 07:40:44.104764 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7sxb" event={"ID":"bb39505c-0ba6-465c-a015-d5be03a5b03e","Type":"ContainerStarted","Data":"6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70"} Nov 23 07:40:45 crc kubenswrapper[4906]: I1123 07:40:45.120780 4906 generic.go:334] "Generic (PLEG): container finished" podID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerID="6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70" exitCode=0 Nov 23 07:40:45 crc kubenswrapper[4906]: I1123 07:40:45.120829 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7sxb" event={"ID":"bb39505c-0ba6-465c-a015-d5be03a5b03e","Type":"ContainerDied","Data":"6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70"} Nov 23 07:40:46 crc kubenswrapper[4906]: I1123 07:40:46.132792 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7sxb" event={"ID":"bb39505c-0ba6-465c-a015-d5be03a5b03e","Type":"ContainerStarted","Data":"db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d"} Nov 23 07:40:46 crc kubenswrapper[4906]: I1123 07:40:46.159738 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d7sxb" podStartSLOduration=2.7102995720000003 podStartE2EDuration="5.15967019s" podCreationTimestamp="2025-11-23 07:40:41 +0000 UTC" firstStartedPulling="2025-11-23 07:40:43.091018957 +0000 UTC m=+3058.604410250" lastFinishedPulling="2025-11-23 07:40:45.540389555 +0000 UTC m=+3061.053780868" observedRunningTime="2025-11-23 07:40:46.156150398 +0000 UTC m=+3061.669541741" watchObservedRunningTime="2025-11-23 07:40:46.15967019 +0000 UTC m=+3061.673061533" Nov 23 07:40:50 crc kubenswrapper[4906]: I1123 07:40:50.946072 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:40:50 crc kubenswrapper[4906]: I1123 07:40:50.946817 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:40:51 crc kubenswrapper[4906]: I1123 07:40:51.848429 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:51 crc kubenswrapper[4906]: I1123 07:40:51.848509 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:51 crc kubenswrapper[4906]: I1123 07:40:51.919415 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:52 crc kubenswrapper[4906]: I1123 07:40:52.266920 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:52 crc kubenswrapper[4906]: I1123 07:40:52.347081 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d7sxb"] Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.208389 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d7sxb" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="registry-server" containerID="cri-o://db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d" gracePeriod=2 Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.753850 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.863871 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjgv8\" (UniqueName: \"kubernetes.io/projected/bb39505c-0ba6-465c-a015-d5be03a5b03e-kube-api-access-qjgv8\") pod \"bb39505c-0ba6-465c-a015-d5be03a5b03e\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.863937 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-catalog-content\") pod \"bb39505c-0ba6-465c-a015-d5be03a5b03e\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.863988 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-utilities\") pod \"bb39505c-0ba6-465c-a015-d5be03a5b03e\" (UID: \"bb39505c-0ba6-465c-a015-d5be03a5b03e\") " Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.865597 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-utilities" (OuterVolumeSpecName: "utilities") pod "bb39505c-0ba6-465c-a015-d5be03a5b03e" (UID: "bb39505c-0ba6-465c-a015-d5be03a5b03e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.871758 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb39505c-0ba6-465c-a015-d5be03a5b03e-kube-api-access-qjgv8" (OuterVolumeSpecName: "kube-api-access-qjgv8") pod "bb39505c-0ba6-465c-a015-d5be03a5b03e" (UID: "bb39505c-0ba6-465c-a015-d5be03a5b03e"). InnerVolumeSpecName "kube-api-access-qjgv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.933144 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb39505c-0ba6-465c-a015-d5be03a5b03e" (UID: "bb39505c-0ba6-465c-a015-d5be03a5b03e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.965945 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjgv8\" (UniqueName: \"kubernetes.io/projected/bb39505c-0ba6-465c-a015-d5be03a5b03e-kube-api-access-qjgv8\") on node \"crc\" DevicePath \"\"" Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.966294 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:40:54 crc kubenswrapper[4906]: I1123 07:40:54.966390 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb39505c-0ba6-465c-a015-d5be03a5b03e-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.223489 4906 generic.go:334] "Generic (PLEG): container finished" podID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerID="db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d" exitCode=0 Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.224053 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7sxb" event={"ID":"bb39505c-0ba6-465c-a015-d5be03a5b03e","Type":"ContainerDied","Data":"db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d"} Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.224156 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7sxb" event={"ID":"bb39505c-0ba6-465c-a015-d5be03a5b03e","Type":"ContainerDied","Data":"a46213a39ef45e026ea3508bcd808b4fb1840e7ff40698ff44585f8adddc8f23"} Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.224186 4906 scope.go:117] "RemoveContainer" containerID="db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.225631 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7sxb" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.257559 4906 scope.go:117] "RemoveContainer" containerID="6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.294802 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d7sxb"] Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.302072 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d7sxb"] Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.310626 4906 scope.go:117] "RemoveContainer" containerID="9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.334430 4906 scope.go:117] "RemoveContainer" containerID="db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d" Nov 23 07:40:55 crc kubenswrapper[4906]: E1123 07:40:55.335136 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d\": container with ID starting with db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d not found: ID does not exist" containerID="db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.335199 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d"} err="failed to get container status \"db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d\": rpc error: code = NotFound desc = could not find container \"db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d\": container with ID starting with db6ca859e462b8fa639a5bbca1f93fea4e24b09d1b874ae24e052c9f87d7303d not found: ID does not exist" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.335236 4906 scope.go:117] "RemoveContainer" containerID="6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70" Nov 23 07:40:55 crc kubenswrapper[4906]: E1123 07:40:55.335822 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70\": container with ID starting with 6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70 not found: ID does not exist" containerID="6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.335883 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70"} err="failed to get container status \"6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70\": rpc error: code = NotFound desc = could not find container \"6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70\": container with ID starting with 6163779819abb681b8d609c081ce893b70fe87dec0e53e1cad5c24ffe2845a70 not found: ID does not exist" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.335922 4906 scope.go:117] "RemoveContainer" containerID="9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0" Nov 23 07:40:55 crc kubenswrapper[4906]: E1123 07:40:55.336768 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0\": container with ID starting with 9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0 not found: ID does not exist" containerID="9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.336811 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0"} err="failed to get container status \"9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0\": rpc error: code = NotFound desc = could not find container \"9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0\": container with ID starting with 9fcb2a095fc0bfee7e96c934dd9a30a290d99962029a19c718ed01f7e282f4f0 not found: ID does not exist" Nov 23 07:40:55 crc kubenswrapper[4906]: I1123 07:40:55.372900 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" path="/var/lib/kubelet/pods/bb39505c-0ba6-465c-a015-d5be03a5b03e/volumes" Nov 23 07:41:20 crc kubenswrapper[4906]: I1123 07:41:20.946423 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:41:20 crc kubenswrapper[4906]: I1123 07:41:20.950062 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:41:50 crc kubenswrapper[4906]: I1123 07:41:50.946711 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:41:50 crc kubenswrapper[4906]: I1123 07:41:50.947252 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:41:50 crc kubenswrapper[4906]: I1123 07:41:50.947393 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:41:50 crc kubenswrapper[4906]: I1123 07:41:50.947985 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:41:50 crc kubenswrapper[4906]: I1123 07:41:50.948036 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" gracePeriod=600 Nov 23 07:41:51 crc kubenswrapper[4906]: E1123 07:41:51.069093 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:41:51 crc kubenswrapper[4906]: I1123 07:41:51.780155 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" exitCode=0 Nov 23 07:41:51 crc kubenswrapper[4906]: I1123 07:41:51.780209 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2"} Nov 23 07:41:51 crc kubenswrapper[4906]: I1123 07:41:51.780260 4906 scope.go:117] "RemoveContainer" containerID="20f1a767b5b092456875c9989d4040177d1080271980f988e48b049e61bf2e20" Nov 23 07:41:51 crc kubenswrapper[4906]: I1123 07:41:51.781194 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:41:51 crc kubenswrapper[4906]: E1123 07:41:51.781459 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:42:04 crc kubenswrapper[4906]: I1123 07:42:04.357348 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:42:04 crc kubenswrapper[4906]: E1123 07:42:04.358289 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:42:16 crc kubenswrapper[4906]: I1123 07:42:16.357170 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:42:16 crc kubenswrapper[4906]: E1123 07:42:16.358521 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:42:30 crc kubenswrapper[4906]: I1123 07:42:30.357865 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:42:30 crc kubenswrapper[4906]: E1123 07:42:30.359358 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:42:42 crc kubenswrapper[4906]: I1123 07:42:42.357042 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:42:42 crc kubenswrapper[4906]: E1123 07:42:42.359036 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:42:56 crc kubenswrapper[4906]: I1123 07:42:56.356966 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:42:56 crc kubenswrapper[4906]: E1123 07:42:56.358333 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:43:08 crc kubenswrapper[4906]: I1123 07:43:08.356738 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:43:08 crc kubenswrapper[4906]: E1123 07:43:08.357489 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:43:21 crc kubenswrapper[4906]: I1123 07:43:21.357025 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:43:21 crc kubenswrapper[4906]: E1123 07:43:21.358376 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:43:35 crc kubenswrapper[4906]: I1123 07:43:35.361170 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:43:35 crc kubenswrapper[4906]: E1123 07:43:35.362101 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:43:49 crc kubenswrapper[4906]: I1123 07:43:49.356579 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:43:49 crc kubenswrapper[4906]: E1123 07:43:49.357596 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:44:02 crc kubenswrapper[4906]: I1123 07:44:02.357053 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:44:02 crc kubenswrapper[4906]: E1123 07:44:02.358027 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:44:15 crc kubenswrapper[4906]: I1123 07:44:15.364918 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:44:15 crc kubenswrapper[4906]: E1123 07:44:15.366269 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:44:27 crc kubenswrapper[4906]: I1123 07:44:27.356186 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:44:27 crc kubenswrapper[4906]: E1123 07:44:27.357138 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:44:30 crc kubenswrapper[4906]: I1123 07:44:30.965820 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-64bmp"] Nov 23 07:44:30 crc kubenswrapper[4906]: E1123 07:44:30.968791 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="registry-server" Nov 23 07:44:30 crc kubenswrapper[4906]: I1123 07:44:30.968903 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="registry-server" Nov 23 07:44:30 crc kubenswrapper[4906]: E1123 07:44:30.969030 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="extract-content" Nov 23 07:44:30 crc kubenswrapper[4906]: I1123 07:44:30.969118 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="extract-content" Nov 23 07:44:30 crc kubenswrapper[4906]: E1123 07:44:30.969203 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="extract-utilities" Nov 23 07:44:30 crc kubenswrapper[4906]: I1123 07:44:30.969278 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="extract-utilities" Nov 23 07:44:30 crc kubenswrapper[4906]: I1123 07:44:30.969546 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb39505c-0ba6-465c-a015-d5be03a5b03e" containerName="registry-server" Nov 23 07:44:30 crc kubenswrapper[4906]: I1123 07:44:30.971190 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:30 crc kubenswrapper[4906]: I1123 07:44:30.987070 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-64bmp"] Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.123261 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zslk\" (UniqueName: \"kubernetes.io/projected/85d67d3a-846a-4a4c-a8b1-aff16cb53154-kube-api-access-9zslk\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.123426 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-utilities\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.123734 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-catalog-content\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.225635 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zslk\" (UniqueName: \"kubernetes.io/projected/85d67d3a-846a-4a4c-a8b1-aff16cb53154-kube-api-access-9zslk\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.225846 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-utilities\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.225933 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-catalog-content\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.227012 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-utilities\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.227119 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-catalog-content\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.256943 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zslk\" (UniqueName: \"kubernetes.io/projected/85d67d3a-846a-4a4c-a8b1-aff16cb53154-kube-api-access-9zslk\") pod \"community-operators-64bmp\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.299032 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.568390 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9xnpc"] Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.570316 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.587299 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xnpc"] Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.734663 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk444\" (UniqueName: \"kubernetes.io/projected/8610e4c8-2052-4bc2-bd94-95058baf69bb-kube-api-access-hk444\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.735044 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-catalog-content\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.735115 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-utilities\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.835711 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-64bmp"] Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.837105 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-catalog-content\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.837198 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-utilities\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.837309 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk444\" (UniqueName: \"kubernetes.io/projected/8610e4c8-2052-4bc2-bd94-95058baf69bb-kube-api-access-hk444\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.839921 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-catalog-content\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.840163 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-utilities\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.864232 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk444\" (UniqueName: \"kubernetes.io/projected/8610e4c8-2052-4bc2-bd94-95058baf69bb-kube-api-access-hk444\") pod \"redhat-marketplace-9xnpc\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.894851 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:31 crc kubenswrapper[4906]: I1123 07:44:31.957599 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64bmp" event={"ID":"85d67d3a-846a-4a4c-a8b1-aff16cb53154","Type":"ContainerStarted","Data":"3675cb6b8d6dda5894f9d5c44c10fda9adc480b6ce555f09cd428a9a14a3ec57"} Nov 23 07:44:32 crc kubenswrapper[4906]: I1123 07:44:32.438252 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xnpc"] Nov 23 07:44:32 crc kubenswrapper[4906]: W1123 07:44:32.448968 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8610e4c8_2052_4bc2_bd94_95058baf69bb.slice/crio-9d8ac6487142f35549fbbfd17398cf19625dc3d3eb32f99e076937d87e9f62ff WatchSource:0}: Error finding container 9d8ac6487142f35549fbbfd17398cf19625dc3d3eb32f99e076937d87e9f62ff: Status 404 returned error can't find the container with id 9d8ac6487142f35549fbbfd17398cf19625dc3d3eb32f99e076937d87e9f62ff Nov 23 07:44:32 crc kubenswrapper[4906]: I1123 07:44:32.969670 4906 generic.go:334] "Generic (PLEG): container finished" podID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerID="fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e" exitCode=0 Nov 23 07:44:32 crc kubenswrapper[4906]: I1123 07:44:32.969814 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64bmp" event={"ID":"85d67d3a-846a-4a4c-a8b1-aff16cb53154","Type":"ContainerDied","Data":"fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e"} Nov 23 07:44:32 crc kubenswrapper[4906]: I1123 07:44:32.971933 4906 generic.go:334] "Generic (PLEG): container finished" podID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerID="c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba" exitCode=0 Nov 23 07:44:32 crc kubenswrapper[4906]: I1123 07:44:32.972002 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xnpc" event={"ID":"8610e4c8-2052-4bc2-bd94-95058baf69bb","Type":"ContainerDied","Data":"c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba"} Nov 23 07:44:32 crc kubenswrapper[4906]: I1123 07:44:32.972044 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xnpc" event={"ID":"8610e4c8-2052-4bc2-bd94-95058baf69bb","Type":"ContainerStarted","Data":"9d8ac6487142f35549fbbfd17398cf19625dc3d3eb32f99e076937d87e9f62ff"} Nov 23 07:44:33 crc kubenswrapper[4906]: I1123 07:44:33.982758 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64bmp" event={"ID":"85d67d3a-846a-4a4c-a8b1-aff16cb53154","Type":"ContainerStarted","Data":"a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670"} Nov 23 07:44:33 crc kubenswrapper[4906]: I1123 07:44:33.987309 4906 generic.go:334] "Generic (PLEG): container finished" podID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerID="45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908" exitCode=0 Nov 23 07:44:33 crc kubenswrapper[4906]: I1123 07:44:33.987490 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xnpc" event={"ID":"8610e4c8-2052-4bc2-bd94-95058baf69bb","Type":"ContainerDied","Data":"45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908"} Nov 23 07:44:35 crc kubenswrapper[4906]: I1123 07:44:35.000969 4906 generic.go:334] "Generic (PLEG): container finished" podID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerID="a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670" exitCode=0 Nov 23 07:44:35 crc kubenswrapper[4906]: I1123 07:44:35.001028 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64bmp" event={"ID":"85d67d3a-846a-4a4c-a8b1-aff16cb53154","Type":"ContainerDied","Data":"a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670"} Nov 23 07:44:35 crc kubenswrapper[4906]: I1123 07:44:35.006448 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xnpc" event={"ID":"8610e4c8-2052-4bc2-bd94-95058baf69bb","Type":"ContainerStarted","Data":"8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010"} Nov 23 07:44:35 crc kubenswrapper[4906]: I1123 07:44:35.073169 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9xnpc" podStartSLOduration=2.6628923970000002 podStartE2EDuration="4.0731422s" podCreationTimestamp="2025-11-23 07:44:31 +0000 UTC" firstStartedPulling="2025-11-23 07:44:32.975957239 +0000 UTC m=+3288.489348542" lastFinishedPulling="2025-11-23 07:44:34.386207022 +0000 UTC m=+3289.899598345" observedRunningTime="2025-11-23 07:44:35.062507358 +0000 UTC m=+3290.575898701" watchObservedRunningTime="2025-11-23 07:44:35.0731422 +0000 UTC m=+3290.586533543" Nov 23 07:44:36 crc kubenswrapper[4906]: I1123 07:44:36.017251 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64bmp" event={"ID":"85d67d3a-846a-4a4c-a8b1-aff16cb53154","Type":"ContainerStarted","Data":"5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d"} Nov 23 07:44:36 crc kubenswrapper[4906]: I1123 07:44:36.049439 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-64bmp" podStartSLOduration=3.608217739 podStartE2EDuration="6.049417529s" podCreationTimestamp="2025-11-23 07:44:30 +0000 UTC" firstStartedPulling="2025-11-23 07:44:32.971649905 +0000 UTC m=+3288.485041258" lastFinishedPulling="2025-11-23 07:44:35.412849755 +0000 UTC m=+3290.926241048" observedRunningTime="2025-11-23 07:44:36.043904583 +0000 UTC m=+3291.557295896" watchObservedRunningTime="2025-11-23 07:44:36.049417529 +0000 UTC m=+3291.562808832" Nov 23 07:44:39 crc kubenswrapper[4906]: I1123 07:44:39.358352 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:44:39 crc kubenswrapper[4906]: E1123 07:44:39.359343 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:44:41 crc kubenswrapper[4906]: I1123 07:44:41.300061 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:41 crc kubenswrapper[4906]: I1123 07:44:41.300117 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:41 crc kubenswrapper[4906]: I1123 07:44:41.371250 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:41 crc kubenswrapper[4906]: I1123 07:44:41.896093 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:41 crc kubenswrapper[4906]: I1123 07:44:41.896181 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:41 crc kubenswrapper[4906]: I1123 07:44:41.945107 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:42 crc kubenswrapper[4906]: I1123 07:44:42.142967 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:42 crc kubenswrapper[4906]: I1123 07:44:42.144468 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:43 crc kubenswrapper[4906]: I1123 07:44:43.019561 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xnpc"] Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.096937 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9xnpc" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="registry-server" containerID="cri-o://8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010" gracePeriod=2 Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.631250 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.767070 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-catalog-content\") pod \"8610e4c8-2052-4bc2-bd94-95058baf69bb\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.768155 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-utilities\") pod \"8610e4c8-2052-4bc2-bd94-95058baf69bb\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.768307 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk444\" (UniqueName: \"kubernetes.io/projected/8610e4c8-2052-4bc2-bd94-95058baf69bb-kube-api-access-hk444\") pod \"8610e4c8-2052-4bc2-bd94-95058baf69bb\" (UID: \"8610e4c8-2052-4bc2-bd94-95058baf69bb\") " Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.769102 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-utilities" (OuterVolumeSpecName: "utilities") pod "8610e4c8-2052-4bc2-bd94-95058baf69bb" (UID: "8610e4c8-2052-4bc2-bd94-95058baf69bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.777338 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8610e4c8-2052-4bc2-bd94-95058baf69bb-kube-api-access-hk444" (OuterVolumeSpecName: "kube-api-access-hk444") pod "8610e4c8-2052-4bc2-bd94-95058baf69bb" (UID: "8610e4c8-2052-4bc2-bd94-95058baf69bb"). InnerVolumeSpecName "kube-api-access-hk444". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.787078 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8610e4c8-2052-4bc2-bd94-95058baf69bb" (UID: "8610e4c8-2052-4bc2-bd94-95058baf69bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.870509 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.870554 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8610e4c8-2052-4bc2-bd94-95058baf69bb-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:44:44 crc kubenswrapper[4906]: I1123 07:44:44.870566 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk444\" (UniqueName: \"kubernetes.io/projected/8610e4c8-2052-4bc2-bd94-95058baf69bb-kube-api-access-hk444\") on node \"crc\" DevicePath \"\"" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.111654 4906 generic.go:334] "Generic (PLEG): container finished" podID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerID="8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010" exitCode=0 Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.111790 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xnpc" event={"ID":"8610e4c8-2052-4bc2-bd94-95058baf69bb","Type":"ContainerDied","Data":"8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010"} Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.111848 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xnpc" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.111923 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xnpc" event={"ID":"8610e4c8-2052-4bc2-bd94-95058baf69bb","Type":"ContainerDied","Data":"9d8ac6487142f35549fbbfd17398cf19625dc3d3eb32f99e076937d87e9f62ff"} Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.111975 4906 scope.go:117] "RemoveContainer" containerID="8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.145479 4906 scope.go:117] "RemoveContainer" containerID="45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.170509 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xnpc"] Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.177926 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xnpc"] Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.196955 4906 scope.go:117] "RemoveContainer" containerID="c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.220697 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-64bmp"] Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.221449 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-64bmp" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="registry-server" containerID="cri-o://5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d" gracePeriod=2 Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.239166 4906 scope.go:117] "RemoveContainer" containerID="8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010" Nov 23 07:44:45 crc kubenswrapper[4906]: E1123 07:44:45.239832 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010\": container with ID starting with 8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010 not found: ID does not exist" containerID="8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.239916 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010"} err="failed to get container status \"8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010\": rpc error: code = NotFound desc = could not find container \"8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010\": container with ID starting with 8a8520b43b8dcd4004ac66b93946650e2bb265e28a3c2439659a74d160588010 not found: ID does not exist" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.239973 4906 scope.go:117] "RemoveContainer" containerID="45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908" Nov 23 07:44:45 crc kubenswrapper[4906]: E1123 07:44:45.240433 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908\": container with ID starting with 45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908 not found: ID does not exist" containerID="45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.240493 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908"} err="failed to get container status \"45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908\": rpc error: code = NotFound desc = could not find container \"45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908\": container with ID starting with 45b17527c9f5bf3469037a446baa4d8d324393ce066422abbef830bc2bb2f908 not found: ID does not exist" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.240526 4906 scope.go:117] "RemoveContainer" containerID="c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba" Nov 23 07:44:45 crc kubenswrapper[4906]: E1123 07:44:45.241120 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba\": container with ID starting with c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba not found: ID does not exist" containerID="c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.241167 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba"} err="failed to get container status \"c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba\": rpc error: code = NotFound desc = could not find container \"c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba\": container with ID starting with c50ab370152de0099dae60af201b0cd00d119829f4812fa1ae8ee82d7fb791ba not found: ID does not exist" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.382416 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" path="/var/lib/kubelet/pods/8610e4c8-2052-4bc2-bd94-95058baf69bb/volumes" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.653351 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.787299 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-catalog-content\") pod \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.787383 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-utilities\") pod \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.787559 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zslk\" (UniqueName: \"kubernetes.io/projected/85d67d3a-846a-4a4c-a8b1-aff16cb53154-kube-api-access-9zslk\") pod \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\" (UID: \"85d67d3a-846a-4a4c-a8b1-aff16cb53154\") " Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.788425 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-utilities" (OuterVolumeSpecName: "utilities") pod "85d67d3a-846a-4a4c-a8b1-aff16cb53154" (UID: "85d67d3a-846a-4a4c-a8b1-aff16cb53154"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.789402 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.800051 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85d67d3a-846a-4a4c-a8b1-aff16cb53154-kube-api-access-9zslk" (OuterVolumeSpecName: "kube-api-access-9zslk") pod "85d67d3a-846a-4a4c-a8b1-aff16cb53154" (UID: "85d67d3a-846a-4a4c-a8b1-aff16cb53154"). InnerVolumeSpecName "kube-api-access-9zslk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.835340 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85d67d3a-846a-4a4c-a8b1-aff16cb53154" (UID: "85d67d3a-846a-4a4c-a8b1-aff16cb53154"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.891344 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d67d3a-846a-4a4c-a8b1-aff16cb53154-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:44:45 crc kubenswrapper[4906]: I1123 07:44:45.891385 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zslk\" (UniqueName: \"kubernetes.io/projected/85d67d3a-846a-4a4c-a8b1-aff16cb53154-kube-api-access-9zslk\") on node \"crc\" DevicePath \"\"" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.129375 4906 generic.go:334] "Generic (PLEG): container finished" podID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerID="5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d" exitCode=0 Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.129541 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64bmp" event={"ID":"85d67d3a-846a-4a4c-a8b1-aff16cb53154","Type":"ContainerDied","Data":"5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d"} Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.129600 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-64bmp" event={"ID":"85d67d3a-846a-4a4c-a8b1-aff16cb53154","Type":"ContainerDied","Data":"3675cb6b8d6dda5894f9d5c44c10fda9adc480b6ce555f09cd428a9a14a3ec57"} Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.129635 4906 scope.go:117] "RemoveContainer" containerID="5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.136223 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-64bmp" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.159319 4906 scope.go:117] "RemoveContainer" containerID="a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.195793 4906 scope.go:117] "RemoveContainer" containerID="fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.198543 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-64bmp"] Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.209205 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-64bmp"] Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.223912 4906 scope.go:117] "RemoveContainer" containerID="5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d" Nov 23 07:44:46 crc kubenswrapper[4906]: E1123 07:44:46.224567 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d\": container with ID starting with 5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d not found: ID does not exist" containerID="5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.224712 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d"} err="failed to get container status \"5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d\": rpc error: code = NotFound desc = could not find container \"5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d\": container with ID starting with 5dbb1b29afa8683eb47b464f078ab6a58e63d6e735f15045ea0a0a0ddae1583d not found: ID does not exist" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.224843 4906 scope.go:117] "RemoveContainer" containerID="a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670" Nov 23 07:44:46 crc kubenswrapper[4906]: E1123 07:44:46.225439 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670\": container with ID starting with a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670 not found: ID does not exist" containerID="a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.225548 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670"} err="failed to get container status \"a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670\": rpc error: code = NotFound desc = could not find container \"a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670\": container with ID starting with a80d1b1aed323b4da1d68b305cf8d93b67ee950db81b042193e650ce63d29670 not found: ID does not exist" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.225643 4906 scope.go:117] "RemoveContainer" containerID="fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e" Nov 23 07:44:46 crc kubenswrapper[4906]: E1123 07:44:46.226144 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e\": container with ID starting with fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e not found: ID does not exist" containerID="fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e" Nov 23 07:44:46 crc kubenswrapper[4906]: I1123 07:44:46.226320 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e"} err="failed to get container status \"fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e\": rpc error: code = NotFound desc = could not find container \"fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e\": container with ID starting with fc31033bfe4d6d091ec0f98d1073b7cb6a023ca52b017a002d4ce2cf753f409e not found: ID does not exist" Nov 23 07:44:47 crc kubenswrapper[4906]: I1123 07:44:47.371562 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" path="/var/lib/kubelet/pods/85d67d3a-846a-4a4c-a8b1-aff16cb53154/volumes" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.439500 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b67fp"] Nov 23 07:44:51 crc kubenswrapper[4906]: E1123 07:44:51.441100 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="extract-content" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441137 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="extract-content" Nov 23 07:44:51 crc kubenswrapper[4906]: E1123 07:44:51.441174 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="registry-server" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441193 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="registry-server" Nov 23 07:44:51 crc kubenswrapper[4906]: E1123 07:44:51.441219 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="extract-utilities" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441238 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="extract-utilities" Nov 23 07:44:51 crc kubenswrapper[4906]: E1123 07:44:51.441265 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="extract-content" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441282 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="extract-content" Nov 23 07:44:51 crc kubenswrapper[4906]: E1123 07:44:51.441321 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="registry-server" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441338 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="registry-server" Nov 23 07:44:51 crc kubenswrapper[4906]: E1123 07:44:51.441375 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="extract-utilities" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441393 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="extract-utilities" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441818 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d67d3a-846a-4a4c-a8b1-aff16cb53154" containerName="registry-server" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.441862 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="8610e4c8-2052-4bc2-bd94-95058baf69bb" containerName="registry-server" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.444926 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.452169 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b67fp"] Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.486751 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-catalog-content\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.486820 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-utilities\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.486857 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhbp9\" (UniqueName: \"kubernetes.io/projected/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-kube-api-access-lhbp9\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.588972 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-utilities\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.589082 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhbp9\" (UniqueName: \"kubernetes.io/projected/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-kube-api-access-lhbp9\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.589231 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-catalog-content\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.590835 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-utilities\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.590908 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-catalog-content\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.620920 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhbp9\" (UniqueName: \"kubernetes.io/projected/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-kube-api-access-lhbp9\") pod \"redhat-operators-b67fp\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:51 crc kubenswrapper[4906]: I1123 07:44:51.778921 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:44:52 crc kubenswrapper[4906]: I1123 07:44:52.272344 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b67fp"] Nov 23 07:44:52 crc kubenswrapper[4906]: I1123 07:44:52.356588 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:44:52 crc kubenswrapper[4906]: E1123 07:44:52.356791 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:44:53 crc kubenswrapper[4906]: I1123 07:44:53.202367 4906 generic.go:334] "Generic (PLEG): container finished" podID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerID="d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09" exitCode=0 Nov 23 07:44:53 crc kubenswrapper[4906]: I1123 07:44:53.202456 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b67fp" event={"ID":"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca","Type":"ContainerDied","Data":"d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09"} Nov 23 07:44:53 crc kubenswrapper[4906]: I1123 07:44:53.204366 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b67fp" event={"ID":"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca","Type":"ContainerStarted","Data":"839f195149ad37fcbdb71a61d73a20b063ac4914e2999678bed5d700ca830fae"} Nov 23 07:44:54 crc kubenswrapper[4906]: I1123 07:44:54.217904 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b67fp" event={"ID":"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca","Type":"ContainerStarted","Data":"335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0"} Nov 23 07:44:55 crc kubenswrapper[4906]: I1123 07:44:55.232345 4906 generic.go:334] "Generic (PLEG): container finished" podID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerID="335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0" exitCode=0 Nov 23 07:44:55 crc kubenswrapper[4906]: I1123 07:44:55.232441 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b67fp" event={"ID":"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca","Type":"ContainerDied","Data":"335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0"} Nov 23 07:44:56 crc kubenswrapper[4906]: I1123 07:44:56.248321 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b67fp" event={"ID":"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca","Type":"ContainerStarted","Data":"dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43"} Nov 23 07:44:56 crc kubenswrapper[4906]: I1123 07:44:56.274094 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b67fp" podStartSLOduration=2.845882785 podStartE2EDuration="5.274074476s" podCreationTimestamp="2025-11-23 07:44:51 +0000 UTC" firstStartedPulling="2025-11-23 07:44:53.206368833 +0000 UTC m=+3308.719760136" lastFinishedPulling="2025-11-23 07:44:55.634560524 +0000 UTC m=+3311.147951827" observedRunningTime="2025-11-23 07:44:56.268769495 +0000 UTC m=+3311.782160808" watchObservedRunningTime="2025-11-23 07:44:56.274074476 +0000 UTC m=+3311.787465779" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.205180 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g"] Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.207076 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.209795 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.214141 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.226941 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g"] Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.359537 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b24sh\" (UniqueName: \"kubernetes.io/projected/f46478a6-465d-4e20-b56a-54904cc46116-kube-api-access-b24sh\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.359624 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f46478a6-465d-4e20-b56a-54904cc46116-secret-volume\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.359840 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f46478a6-465d-4e20-b56a-54904cc46116-config-volume\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.462057 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b24sh\" (UniqueName: \"kubernetes.io/projected/f46478a6-465d-4e20-b56a-54904cc46116-kube-api-access-b24sh\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.462120 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f46478a6-465d-4e20-b56a-54904cc46116-secret-volume\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.462170 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f46478a6-465d-4e20-b56a-54904cc46116-config-volume\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.463244 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f46478a6-465d-4e20-b56a-54904cc46116-config-volume\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.473095 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f46478a6-465d-4e20-b56a-54904cc46116-secret-volume\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.493494 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b24sh\" (UniqueName: \"kubernetes.io/projected/f46478a6-465d-4e20-b56a-54904cc46116-kube-api-access-b24sh\") pod \"collect-profiles-29398065-2ct6g\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:00 crc kubenswrapper[4906]: I1123 07:45:00.533132 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:01 crc kubenswrapper[4906]: I1123 07:45:01.025877 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g"] Nov 23 07:45:01 crc kubenswrapper[4906]: I1123 07:45:01.298540 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" event={"ID":"f46478a6-465d-4e20-b56a-54904cc46116","Type":"ContainerStarted","Data":"b2f9731718c7c464e81bb2a2b0dfd45be0c1218fa1458f99d657e33d28cf414e"} Nov 23 07:45:01 crc kubenswrapper[4906]: I1123 07:45:01.299075 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" event={"ID":"f46478a6-465d-4e20-b56a-54904cc46116","Type":"ContainerStarted","Data":"ff0860ceaaed47d52970b67b7075ef2080bd73fd2f3031b6eecff4da5a3b1deb"} Nov 23 07:45:01 crc kubenswrapper[4906]: I1123 07:45:01.353279 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" podStartSLOduration=1.353247296 podStartE2EDuration="1.353247296s" podCreationTimestamp="2025-11-23 07:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:45:01.313704807 +0000 UTC m=+3316.827096110" watchObservedRunningTime="2025-11-23 07:45:01.353247296 +0000 UTC m=+3316.866638589" Nov 23 07:45:01 crc kubenswrapper[4906]: I1123 07:45:01.779799 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:45:01 crc kubenswrapper[4906]: I1123 07:45:01.779866 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:45:02 crc kubenswrapper[4906]: I1123 07:45:02.312989 4906 generic.go:334] "Generic (PLEG): container finished" podID="f46478a6-465d-4e20-b56a-54904cc46116" containerID="b2f9731718c7c464e81bb2a2b0dfd45be0c1218fa1458f99d657e33d28cf414e" exitCode=0 Nov 23 07:45:02 crc kubenswrapper[4906]: I1123 07:45:02.313071 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" event={"ID":"f46478a6-465d-4e20-b56a-54904cc46116","Type":"ContainerDied","Data":"b2f9731718c7c464e81bb2a2b0dfd45be0c1218fa1458f99d657e33d28cf414e"} Nov 23 07:45:02 crc kubenswrapper[4906]: I1123 07:45:02.853849 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b67fp" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="registry-server" probeResult="failure" output=< Nov 23 07:45:02 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 07:45:02 crc kubenswrapper[4906]: > Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.587790 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.721929 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f46478a6-465d-4e20-b56a-54904cc46116-secret-volume\") pod \"f46478a6-465d-4e20-b56a-54904cc46116\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.722213 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f46478a6-465d-4e20-b56a-54904cc46116-config-volume\") pod \"f46478a6-465d-4e20-b56a-54904cc46116\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.722277 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b24sh\" (UniqueName: \"kubernetes.io/projected/f46478a6-465d-4e20-b56a-54904cc46116-kube-api-access-b24sh\") pod \"f46478a6-465d-4e20-b56a-54904cc46116\" (UID: \"f46478a6-465d-4e20-b56a-54904cc46116\") " Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.723221 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f46478a6-465d-4e20-b56a-54904cc46116-config-volume" (OuterVolumeSpecName: "config-volume") pod "f46478a6-465d-4e20-b56a-54904cc46116" (UID: "f46478a6-465d-4e20-b56a-54904cc46116"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.729467 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f46478a6-465d-4e20-b56a-54904cc46116-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f46478a6-465d-4e20-b56a-54904cc46116" (UID: "f46478a6-465d-4e20-b56a-54904cc46116"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.730792 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f46478a6-465d-4e20-b56a-54904cc46116-kube-api-access-b24sh" (OuterVolumeSpecName: "kube-api-access-b24sh") pod "f46478a6-465d-4e20-b56a-54904cc46116" (UID: "f46478a6-465d-4e20-b56a-54904cc46116"). InnerVolumeSpecName "kube-api-access-b24sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.824221 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f46478a6-465d-4e20-b56a-54904cc46116-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.824276 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b24sh\" (UniqueName: \"kubernetes.io/projected/f46478a6-465d-4e20-b56a-54904cc46116-kube-api-access-b24sh\") on node \"crc\" DevicePath \"\"" Nov 23 07:45:03 crc kubenswrapper[4906]: I1123 07:45:03.824289 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f46478a6-465d-4e20-b56a-54904cc46116-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:45:04 crc kubenswrapper[4906]: I1123 07:45:04.331244 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" event={"ID":"f46478a6-465d-4e20-b56a-54904cc46116","Type":"ContainerDied","Data":"ff0860ceaaed47d52970b67b7075ef2080bd73fd2f3031b6eecff4da5a3b1deb"} Nov 23 07:45:04 crc kubenswrapper[4906]: I1123 07:45:04.331298 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff0860ceaaed47d52970b67b7075ef2080bd73fd2f3031b6eecff4da5a3b1deb" Nov 23 07:45:04 crc kubenswrapper[4906]: I1123 07:45:04.331318 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g" Nov 23 07:45:04 crc kubenswrapper[4906]: I1123 07:45:04.407384 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6"] Nov 23 07:45:04 crc kubenswrapper[4906]: I1123 07:45:04.412813 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-szcs6"] Nov 23 07:45:05 crc kubenswrapper[4906]: I1123 07:45:05.378333 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2dfef9e-bfcd-4419-aa27-773878c02e80" path="/var/lib/kubelet/pods/c2dfef9e-bfcd-4419-aa27-773878c02e80/volumes" Nov 23 07:45:07 crc kubenswrapper[4906]: I1123 07:45:07.356181 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:45:07 crc kubenswrapper[4906]: E1123 07:45:07.357397 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:45:11 crc kubenswrapper[4906]: I1123 07:45:11.858429 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:45:11 crc kubenswrapper[4906]: I1123 07:45:11.951900 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:45:12 crc kubenswrapper[4906]: I1123 07:45:12.118296 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b67fp"] Nov 23 07:45:13 crc kubenswrapper[4906]: I1123 07:45:13.447505 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b67fp" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="registry-server" containerID="cri-o://dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43" gracePeriod=2 Nov 23 07:45:13 crc kubenswrapper[4906]: I1123 07:45:13.899797 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.005932 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-catalog-content\") pod \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.006641 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-utilities\") pod \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.006765 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhbp9\" (UniqueName: \"kubernetes.io/projected/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-kube-api-access-lhbp9\") pod \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\" (UID: \"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca\") " Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.009043 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-utilities" (OuterVolumeSpecName: "utilities") pod "2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" (UID: "2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.015792 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-kube-api-access-lhbp9" (OuterVolumeSpecName: "kube-api-access-lhbp9") pod "2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" (UID: "2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca"). InnerVolumeSpecName "kube-api-access-lhbp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.109362 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.109411 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhbp9\" (UniqueName: \"kubernetes.io/projected/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-kube-api-access-lhbp9\") on node \"crc\" DevicePath \"\"" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.121036 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" (UID: "2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.211514 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.460106 4906 generic.go:334] "Generic (PLEG): container finished" podID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerID="dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43" exitCode=0 Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.460167 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b67fp" event={"ID":"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca","Type":"ContainerDied","Data":"dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43"} Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.460210 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b67fp" event={"ID":"2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca","Type":"ContainerDied","Data":"839f195149ad37fcbdb71a61d73a20b063ac4914e2999678bed5d700ca830fae"} Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.460231 4906 scope.go:117] "RemoveContainer" containerID="dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.460252 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b67fp" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.504502 4906 scope.go:117] "RemoveContainer" containerID="335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.511672 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b67fp"] Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.518213 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b67fp"] Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.534539 4906 scope.go:117] "RemoveContainer" containerID="d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.576070 4906 scope.go:117] "RemoveContainer" containerID="dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43" Nov 23 07:45:14 crc kubenswrapper[4906]: E1123 07:45:14.576563 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43\": container with ID starting with dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43 not found: ID does not exist" containerID="dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.576612 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43"} err="failed to get container status \"dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43\": rpc error: code = NotFound desc = could not find container \"dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43\": container with ID starting with dda1a09fd52036fffc55654a5c54095d305f94d591a7ab2aaf2bdb46b67a8c43 not found: ID does not exist" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.576653 4906 scope.go:117] "RemoveContainer" containerID="335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0" Nov 23 07:45:14 crc kubenswrapper[4906]: E1123 07:45:14.577237 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0\": container with ID starting with 335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0 not found: ID does not exist" containerID="335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.577278 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0"} err="failed to get container status \"335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0\": rpc error: code = NotFound desc = could not find container \"335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0\": container with ID starting with 335574186d09ebb01d3faf1efc3462181add0b8567598f328cdb10c86a771bc0 not found: ID does not exist" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.577301 4906 scope.go:117] "RemoveContainer" containerID="d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09" Nov 23 07:45:14 crc kubenswrapper[4906]: E1123 07:45:14.577871 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09\": container with ID starting with d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09 not found: ID does not exist" containerID="d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09" Nov 23 07:45:14 crc kubenswrapper[4906]: I1123 07:45:14.577906 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09"} err="failed to get container status \"d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09\": rpc error: code = NotFound desc = could not find container \"d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09\": container with ID starting with d9fa674288d1a261c236ea83f5adb7483feadeaedf10d35d58db1ae20f345b09 not found: ID does not exist" Nov 23 07:45:15 crc kubenswrapper[4906]: I1123 07:45:15.375839 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" path="/var/lib/kubelet/pods/2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca/volumes" Nov 23 07:45:19 crc kubenswrapper[4906]: I1123 07:45:19.358276 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:45:19 crc kubenswrapper[4906]: E1123 07:45:19.361629 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:45:32 crc kubenswrapper[4906]: I1123 07:45:32.356485 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:45:32 crc kubenswrapper[4906]: E1123 07:45:32.357051 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:45:44 crc kubenswrapper[4906]: I1123 07:45:44.356482 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:45:44 crc kubenswrapper[4906]: E1123 07:45:44.357361 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:45:55 crc kubenswrapper[4906]: I1123 07:45:55.370517 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:45:55 crc kubenswrapper[4906]: E1123 07:45:55.371856 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:45:57 crc kubenswrapper[4906]: I1123 07:45:57.244221 4906 scope.go:117] "RemoveContainer" containerID="67f58f916b7ffbc0e149b0c1d2b10f3e0edf3c34d857a2b184fc78da0a740b29" Nov 23 07:46:09 crc kubenswrapper[4906]: I1123 07:46:09.357370 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:46:09 crc kubenswrapper[4906]: E1123 07:46:09.358726 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:46:24 crc kubenswrapper[4906]: I1123 07:46:24.357388 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:46:24 crc kubenswrapper[4906]: E1123 07:46:24.358726 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:46:37 crc kubenswrapper[4906]: I1123 07:46:37.357062 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:46:37 crc kubenswrapper[4906]: E1123 07:46:37.358198 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:46:48 crc kubenswrapper[4906]: I1123 07:46:48.356599 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:46:48 crc kubenswrapper[4906]: E1123 07:46:48.357568 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:47:01 crc kubenswrapper[4906]: I1123 07:47:01.356653 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:47:02 crc kubenswrapper[4906]: I1123 07:47:02.507809 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"8d24c84e811d593e2953cc78677e27a38b5b7e6db7dc989ff974e75e62265b36"} Nov 23 07:49:20 crc kubenswrapper[4906]: I1123 07:49:20.945785 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:49:20 crc kubenswrapper[4906]: I1123 07:49:20.946363 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:49:50 crc kubenswrapper[4906]: I1123 07:49:50.946123 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:49:50 crc kubenswrapper[4906]: I1123 07:49:50.947240 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:50:20 crc kubenswrapper[4906]: I1123 07:50:20.945943 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:50:20 crc kubenswrapper[4906]: I1123 07:50:20.946532 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:50:20 crc kubenswrapper[4906]: I1123 07:50:20.946587 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:50:20 crc kubenswrapper[4906]: I1123 07:50:20.947375 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d24c84e811d593e2953cc78677e27a38b5b7e6db7dc989ff974e75e62265b36"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:50:20 crc kubenswrapper[4906]: I1123 07:50:20.947456 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://8d24c84e811d593e2953cc78677e27a38b5b7e6db7dc989ff974e75e62265b36" gracePeriod=600 Nov 23 07:50:21 crc kubenswrapper[4906]: I1123 07:50:21.599491 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="8d24c84e811d593e2953cc78677e27a38b5b7e6db7dc989ff974e75e62265b36" exitCode=0 Nov 23 07:50:21 crc kubenswrapper[4906]: I1123 07:50:21.599596 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"8d24c84e811d593e2953cc78677e27a38b5b7e6db7dc989ff974e75e62265b36"} Nov 23 07:50:21 crc kubenswrapper[4906]: I1123 07:50:21.600368 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51"} Nov 23 07:50:21 crc kubenswrapper[4906]: I1123 07:50:21.600409 4906 scope.go:117] "RemoveContainer" containerID="97bc1b4b0cdc76015bbdaddb71c15a13eaffd9220013fe8c556ea4d25620f0e2" Nov 23 07:52:50 crc kubenswrapper[4906]: I1123 07:52:50.946164 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:52:50 crc kubenswrapper[4906]: I1123 07:52:50.947008 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:53:20 crc kubenswrapper[4906]: I1123 07:53:20.946345 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:53:20 crc kubenswrapper[4906]: I1123 07:53:20.947887 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:53:50 crc kubenswrapper[4906]: I1123 07:53:50.946250 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:53:50 crc kubenswrapper[4906]: I1123 07:53:50.947315 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:53:50 crc kubenswrapper[4906]: I1123 07:53:50.947406 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 07:53:50 crc kubenswrapper[4906]: I1123 07:53:50.948650 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:53:50 crc kubenswrapper[4906]: I1123 07:53:50.948794 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" gracePeriod=600 Nov 23 07:53:51 crc kubenswrapper[4906]: E1123 07:53:51.080773 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:53:51 crc kubenswrapper[4906]: I1123 07:53:51.715202 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" exitCode=0 Nov 23 07:53:51 crc kubenswrapper[4906]: I1123 07:53:51.715290 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51"} Nov 23 07:53:51 crc kubenswrapper[4906]: I1123 07:53:51.715381 4906 scope.go:117] "RemoveContainer" containerID="8d24c84e811d593e2953cc78677e27a38b5b7e6db7dc989ff974e75e62265b36" Nov 23 07:53:51 crc kubenswrapper[4906]: I1123 07:53:51.716636 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:53:51 crc kubenswrapper[4906]: E1123 07:53:51.717157 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:54:05 crc kubenswrapper[4906]: I1123 07:54:05.365411 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:54:05 crc kubenswrapper[4906]: E1123 07:54:05.366753 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:54:17 crc kubenswrapper[4906]: I1123 07:54:17.357732 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:54:17 crc kubenswrapper[4906]: E1123 07:54:17.359244 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:54:29 crc kubenswrapper[4906]: I1123 07:54:29.357611 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:54:29 crc kubenswrapper[4906]: E1123 07:54:29.358998 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.867525 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xvk22"] Nov 23 07:54:33 crc kubenswrapper[4906]: E1123 07:54:33.868490 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f46478a6-465d-4e20-b56a-54904cc46116" containerName="collect-profiles" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.868505 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f46478a6-465d-4e20-b56a-54904cc46116" containerName="collect-profiles" Nov 23 07:54:33 crc kubenswrapper[4906]: E1123 07:54:33.868524 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="extract-utilities" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.868530 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="extract-utilities" Nov 23 07:54:33 crc kubenswrapper[4906]: E1123 07:54:33.868540 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="registry-server" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.868547 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="registry-server" Nov 23 07:54:33 crc kubenswrapper[4906]: E1123 07:54:33.868557 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="extract-content" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.868563 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="extract-content" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.868722 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="f46478a6-465d-4e20-b56a-54904cc46116" containerName="collect-profiles" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.868746 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b47c2a4-9043-4714-b0b2-9f6c9b5d79ca" containerName="registry-server" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.869785 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.914861 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvk22"] Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.940868 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-catalog-content\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.941104 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xjxv\" (UniqueName: \"kubernetes.io/projected/e8effd89-14c9-45e0-9c13-6ee58c5d80af-kube-api-access-8xjxv\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:33 crc kubenswrapper[4906]: I1123 07:54:33.941212 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-utilities\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.043652 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xjxv\" (UniqueName: \"kubernetes.io/projected/e8effd89-14c9-45e0-9c13-6ee58c5d80af-kube-api-access-8xjxv\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.043771 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-utilities\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.043894 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-catalog-content\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.044372 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-utilities\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.044552 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-catalog-content\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.069538 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xjxv\" (UniqueName: \"kubernetes.io/projected/e8effd89-14c9-45e0-9c13-6ee58c5d80af-kube-api-access-8xjxv\") pod \"community-operators-xvk22\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.216145 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:34 crc kubenswrapper[4906]: I1123 07:54:34.573169 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvk22"] Nov 23 07:54:35 crc kubenswrapper[4906]: I1123 07:54:35.185460 4906 generic.go:334] "Generic (PLEG): container finished" podID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerID="1e02e4838a754b86154bf6212219c7d23c6cd07ca317e019d6ca3cb47876cdaf" exitCode=0 Nov 23 07:54:35 crc kubenswrapper[4906]: I1123 07:54:35.185522 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvk22" event={"ID":"e8effd89-14c9-45e0-9c13-6ee58c5d80af","Type":"ContainerDied","Data":"1e02e4838a754b86154bf6212219c7d23c6cd07ca317e019d6ca3cb47876cdaf"} Nov 23 07:54:35 crc kubenswrapper[4906]: I1123 07:54:35.185556 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvk22" event={"ID":"e8effd89-14c9-45e0-9c13-6ee58c5d80af","Type":"ContainerStarted","Data":"7fa344beb6d12455a2b53d3cffb3989d8bd88b117bac660e0fa407417aa54f7b"} Nov 23 07:54:35 crc kubenswrapper[4906]: I1123 07:54:35.189302 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:54:36 crc kubenswrapper[4906]: I1123 07:54:36.193565 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvk22" event={"ID":"e8effd89-14c9-45e0-9c13-6ee58c5d80af","Type":"ContainerStarted","Data":"a755255fff7ae443ab7c7bdfe587c3014f547385266101a3bb3e629582180cb7"} Nov 23 07:54:37 crc kubenswrapper[4906]: I1123 07:54:37.210173 4906 generic.go:334] "Generic (PLEG): container finished" podID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerID="a755255fff7ae443ab7c7bdfe587c3014f547385266101a3bb3e629582180cb7" exitCode=0 Nov 23 07:54:37 crc kubenswrapper[4906]: I1123 07:54:37.210272 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvk22" event={"ID":"e8effd89-14c9-45e0-9c13-6ee58c5d80af","Type":"ContainerDied","Data":"a755255fff7ae443ab7c7bdfe587c3014f547385266101a3bb3e629582180cb7"} Nov 23 07:54:37 crc kubenswrapper[4906]: I1123 07:54:37.210585 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvk22" event={"ID":"e8effd89-14c9-45e0-9c13-6ee58c5d80af","Type":"ContainerStarted","Data":"41a8bcdd6ef8d6168db951a4515f0e58d0247a4dae56d6ca0d85cd3d6c9d30b2"} Nov 23 07:54:37 crc kubenswrapper[4906]: I1123 07:54:37.241338 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xvk22" podStartSLOduration=2.838127255 podStartE2EDuration="4.241321447s" podCreationTimestamp="2025-11-23 07:54:33 +0000 UTC" firstStartedPulling="2025-11-23 07:54:35.189025821 +0000 UTC m=+3890.702417124" lastFinishedPulling="2025-11-23 07:54:36.592220013 +0000 UTC m=+3892.105611316" observedRunningTime="2025-11-23 07:54:37.239175111 +0000 UTC m=+3892.752566424" watchObservedRunningTime="2025-11-23 07:54:37.241321447 +0000 UTC m=+3892.754712750" Nov 23 07:54:44 crc kubenswrapper[4906]: I1123 07:54:44.216918 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:44 crc kubenswrapper[4906]: I1123 07:54:44.217427 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:44 crc kubenswrapper[4906]: I1123 07:54:44.293087 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:44 crc kubenswrapper[4906]: I1123 07:54:44.356899 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:54:44 crc kubenswrapper[4906]: E1123 07:54:44.357319 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:54:44 crc kubenswrapper[4906]: I1123 07:54:44.380523 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:44 crc kubenswrapper[4906]: I1123 07:54:44.543164 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvk22"] Nov 23 07:54:46 crc kubenswrapper[4906]: I1123 07:54:46.299045 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xvk22" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="registry-server" containerID="cri-o://41a8bcdd6ef8d6168db951a4515f0e58d0247a4dae56d6ca0d85cd3d6c9d30b2" gracePeriod=2 Nov 23 07:54:46 crc kubenswrapper[4906]: I1123 07:54:46.952866 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bvz9x"] Nov 23 07:54:46 crc kubenswrapper[4906]: I1123 07:54:46.961185 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:46 crc kubenswrapper[4906]: I1123 07:54:46.964368 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-utilities\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:46 crc kubenswrapper[4906]: I1123 07:54:46.964424 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frgs5\" (UniqueName: \"kubernetes.io/projected/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-kube-api-access-frgs5\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:46 crc kubenswrapper[4906]: I1123 07:54:46.964492 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-catalog-content\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:46 crc kubenswrapper[4906]: I1123 07:54:46.966638 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bvz9x"] Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.066217 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-utilities\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.066599 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frgs5\" (UniqueName: \"kubernetes.io/projected/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-kube-api-access-frgs5\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.067257 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-catalog-content\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.067496 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-utilities\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.067673 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-catalog-content\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.088400 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frgs5\" (UniqueName: \"kubernetes.io/projected/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-kube-api-access-frgs5\") pod \"redhat-marketplace-bvz9x\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.311049 4906 generic.go:334] "Generic (PLEG): container finished" podID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerID="41a8bcdd6ef8d6168db951a4515f0e58d0247a4dae56d6ca0d85cd3d6c9d30b2" exitCode=0 Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.311128 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvk22" event={"ID":"e8effd89-14c9-45e0-9c13-6ee58c5d80af","Type":"ContainerDied","Data":"41a8bcdd6ef8d6168db951a4515f0e58d0247a4dae56d6ca0d85cd3d6c9d30b2"} Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.311184 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvk22" event={"ID":"e8effd89-14c9-45e0-9c13-6ee58c5d80af","Type":"ContainerDied","Data":"7fa344beb6d12455a2b53d3cffb3989d8bd88b117bac660e0fa407417aa54f7b"} Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.311214 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fa344beb6d12455a2b53d3cffb3989d8bd88b117bac660e0fa407417aa54f7b" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.317763 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.319243 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.476787 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-utilities\") pod \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.477358 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xjxv\" (UniqueName: \"kubernetes.io/projected/e8effd89-14c9-45e0-9c13-6ee58c5d80af-kube-api-access-8xjxv\") pod \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.477403 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-catalog-content\") pod \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\" (UID: \"e8effd89-14c9-45e0-9c13-6ee58c5d80af\") " Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.478111 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-utilities" (OuterVolumeSpecName: "utilities") pod "e8effd89-14c9-45e0-9c13-6ee58c5d80af" (UID: "e8effd89-14c9-45e0-9c13-6ee58c5d80af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.489548 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8effd89-14c9-45e0-9c13-6ee58c5d80af-kube-api-access-8xjxv" (OuterVolumeSpecName: "kube-api-access-8xjxv") pod "e8effd89-14c9-45e0-9c13-6ee58c5d80af" (UID: "e8effd89-14c9-45e0-9c13-6ee58c5d80af"). InnerVolumeSpecName "kube-api-access-8xjxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.547537 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8effd89-14c9-45e0-9c13-6ee58c5d80af" (UID: "e8effd89-14c9-45e0-9c13-6ee58c5d80af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.578852 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xjxv\" (UniqueName: \"kubernetes.io/projected/e8effd89-14c9-45e0-9c13-6ee58c5d80af-kube-api-access-8xjxv\") on node \"crc\" DevicePath \"\"" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.578882 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.578893 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8effd89-14c9-45e0-9c13-6ee58c5d80af-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:54:47 crc kubenswrapper[4906]: I1123 07:54:47.810773 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bvz9x"] Nov 23 07:54:48 crc kubenswrapper[4906]: I1123 07:54:48.330629 4906 generic.go:334] "Generic (PLEG): container finished" podID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerID="82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde" exitCode=0 Nov 23 07:54:48 crc kubenswrapper[4906]: I1123 07:54:48.330741 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bvz9x" event={"ID":"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73","Type":"ContainerDied","Data":"82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde"} Nov 23 07:54:48 crc kubenswrapper[4906]: I1123 07:54:48.331406 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvk22" Nov 23 07:54:48 crc kubenswrapper[4906]: I1123 07:54:48.331403 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bvz9x" event={"ID":"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73","Type":"ContainerStarted","Data":"bd17600e2e9f2db0c3a4c5f2d19ee81605261c19d330092e50b86bb60af38094"} Nov 23 07:54:48 crc kubenswrapper[4906]: I1123 07:54:48.383515 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvk22"] Nov 23 07:54:48 crc kubenswrapper[4906]: I1123 07:54:48.390471 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xvk22"] Nov 23 07:54:49 crc kubenswrapper[4906]: I1123 07:54:49.344211 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bvz9x" event={"ID":"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73","Type":"ContainerStarted","Data":"7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9"} Nov 23 07:54:49 crc kubenswrapper[4906]: I1123 07:54:49.375957 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" path="/var/lib/kubelet/pods/e8effd89-14c9-45e0-9c13-6ee58c5d80af/volumes" Nov 23 07:54:50 crc kubenswrapper[4906]: I1123 07:54:50.367468 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bvz9x" event={"ID":"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73","Type":"ContainerDied","Data":"7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9"} Nov 23 07:54:50 crc kubenswrapper[4906]: I1123 07:54:50.367472 4906 generic.go:334] "Generic (PLEG): container finished" podID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerID="7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9" exitCode=0 Nov 23 07:54:51 crc kubenswrapper[4906]: I1123 07:54:51.381399 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bvz9x" event={"ID":"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73","Type":"ContainerStarted","Data":"0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63"} Nov 23 07:54:51 crc kubenswrapper[4906]: I1123 07:54:51.430273 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bvz9x" podStartSLOduration=2.960073792 podStartE2EDuration="5.430242658s" podCreationTimestamp="2025-11-23 07:54:46 +0000 UTC" firstStartedPulling="2025-11-23 07:54:48.333609715 +0000 UTC m=+3903.847001028" lastFinishedPulling="2025-11-23 07:54:50.803778551 +0000 UTC m=+3906.317169894" observedRunningTime="2025-11-23 07:54:51.416439443 +0000 UTC m=+3906.929830776" watchObservedRunningTime="2025-11-23 07:54:51.430242658 +0000 UTC m=+3906.943633971" Nov 23 07:54:55 crc kubenswrapper[4906]: I1123 07:54:55.366152 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:54:55 crc kubenswrapper[4906]: E1123 07:54:55.367535 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:54:57 crc kubenswrapper[4906]: I1123 07:54:57.318649 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:57 crc kubenswrapper[4906]: I1123 07:54:57.319317 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:57 crc kubenswrapper[4906]: I1123 07:54:57.406855 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:57 crc kubenswrapper[4906]: I1123 07:54:57.520334 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:54:57 crc kubenswrapper[4906]: I1123 07:54:57.666794 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bvz9x"] Nov 23 07:54:59 crc kubenswrapper[4906]: I1123 07:54:59.475402 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bvz9x" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="registry-server" containerID="cri-o://0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63" gracePeriod=2 Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.170135 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.255202 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-utilities\") pod \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.255322 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-catalog-content\") pod \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.255364 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frgs5\" (UniqueName: \"kubernetes.io/projected/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-kube-api-access-frgs5\") pod \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\" (UID: \"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73\") " Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.258006 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-utilities" (OuterVolumeSpecName: "utilities") pod "c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" (UID: "c9ddc4c1-46b8-489d-a0db-a6bacde0cb73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.264158 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-kube-api-access-frgs5" (OuterVolumeSpecName: "kube-api-access-frgs5") pod "c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" (UID: "c9ddc4c1-46b8-489d-a0db-a6bacde0cb73"). InnerVolumeSpecName "kube-api-access-frgs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.275522 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" (UID: "c9ddc4c1-46b8-489d-a0db-a6bacde0cb73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.356875 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.357386 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.357473 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frgs5\" (UniqueName: \"kubernetes.io/projected/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73-kube-api-access-frgs5\") on node \"crc\" DevicePath \"\"" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.489657 4906 generic.go:334] "Generic (PLEG): container finished" podID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerID="0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63" exitCode=0 Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.489735 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bvz9x" event={"ID":"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73","Type":"ContainerDied","Data":"0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63"} Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.489803 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bvz9x" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.489826 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bvz9x" event={"ID":"c9ddc4c1-46b8-489d-a0db-a6bacde0cb73","Type":"ContainerDied","Data":"bd17600e2e9f2db0c3a4c5f2d19ee81605261c19d330092e50b86bb60af38094"} Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.489864 4906 scope.go:117] "RemoveContainer" containerID="0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.532527 4906 scope.go:117] "RemoveContainer" containerID="7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.539784 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bvz9x"] Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.545081 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bvz9x"] Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.555744 4906 scope.go:117] "RemoveContainer" containerID="82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.592594 4906 scope.go:117] "RemoveContainer" containerID="0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63" Nov 23 07:55:00 crc kubenswrapper[4906]: E1123 07:55:00.595219 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63\": container with ID starting with 0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63 not found: ID does not exist" containerID="0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.595282 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63"} err="failed to get container status \"0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63\": rpc error: code = NotFound desc = could not find container \"0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63\": container with ID starting with 0adcbcb66c3e9408247a518176156d523f7702b1e4f8dd3b2750bd1d1d2d6d63 not found: ID does not exist" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.595323 4906 scope.go:117] "RemoveContainer" containerID="7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9" Nov 23 07:55:00 crc kubenswrapper[4906]: E1123 07:55:00.596446 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9\": container with ID starting with 7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9 not found: ID does not exist" containerID="7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.596492 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9"} err="failed to get container status \"7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9\": rpc error: code = NotFound desc = could not find container \"7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9\": container with ID starting with 7cc6be1391d19636bf45771430f00b23ba2e850b59360cb1daf44a35e914e1e9 not found: ID does not exist" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.596523 4906 scope.go:117] "RemoveContainer" containerID="82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde" Nov 23 07:55:00 crc kubenswrapper[4906]: E1123 07:55:00.600059 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde\": container with ID starting with 82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde not found: ID does not exist" containerID="82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde" Nov 23 07:55:00 crc kubenswrapper[4906]: I1123 07:55:00.600120 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde"} err="failed to get container status \"82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde\": rpc error: code = NotFound desc = could not find container \"82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde\": container with ID starting with 82b81c4a0c6889b617e04839ad086ca51689e76fa8ab94f474aca336d9571fde not found: ID does not exist" Nov 23 07:55:01 crc kubenswrapper[4906]: I1123 07:55:01.374647 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" path="/var/lib/kubelet/pods/c9ddc4c1-46b8-489d-a0db-a6bacde0cb73/volumes" Nov 23 07:55:07 crc kubenswrapper[4906]: I1123 07:55:07.357404 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:55:07 crc kubenswrapper[4906]: E1123 07:55:07.358626 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:55:21 crc kubenswrapper[4906]: I1123 07:55:21.360475 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:55:21 crc kubenswrapper[4906]: E1123 07:55:21.363024 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:55:32 crc kubenswrapper[4906]: I1123 07:55:32.358762 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:55:32 crc kubenswrapper[4906]: E1123 07:55:32.360395 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.590941 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5kqx"] Nov 23 07:55:43 crc kubenswrapper[4906]: E1123 07:55:43.592148 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="extract-utilities" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592167 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="extract-utilities" Nov 23 07:55:43 crc kubenswrapper[4906]: E1123 07:55:43.592209 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="registry-server" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592217 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="registry-server" Nov 23 07:55:43 crc kubenswrapper[4906]: E1123 07:55:43.592234 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="extract-content" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592245 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="extract-content" Nov 23 07:55:43 crc kubenswrapper[4906]: E1123 07:55:43.592282 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="extract-content" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592292 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="extract-content" Nov 23 07:55:43 crc kubenswrapper[4906]: E1123 07:55:43.592302 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="extract-utilities" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592311 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="extract-utilities" Nov 23 07:55:43 crc kubenswrapper[4906]: E1123 07:55:43.592327 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="registry-server" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592335 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="registry-server" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592524 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ddc4c1-46b8-489d-a0db-a6bacde0cb73" containerName="registry-server" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.592558 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8effd89-14c9-45e0-9c13-6ee58c5d80af" containerName="registry-server" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.593879 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.607996 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5kqx"] Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.764051 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-catalog-content\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.764134 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgm6t\" (UniqueName: \"kubernetes.io/projected/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-kube-api-access-kgm6t\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.764185 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-utilities\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.865514 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-utilities\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.865602 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-catalog-content\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.865706 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgm6t\" (UniqueName: \"kubernetes.io/projected/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-kube-api-access-kgm6t\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.866437 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-utilities\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.866512 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-catalog-content\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.884455 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgm6t\" (UniqueName: \"kubernetes.io/projected/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-kube-api-access-kgm6t\") pod \"redhat-operators-g5kqx\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:43 crc kubenswrapper[4906]: I1123 07:55:43.942615 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:44 crc kubenswrapper[4906]: I1123 07:55:44.393052 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5kqx"] Nov 23 07:55:44 crc kubenswrapper[4906]: I1123 07:55:44.931584 4906 generic.go:334] "Generic (PLEG): container finished" podID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerID="0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e" exitCode=0 Nov 23 07:55:44 crc kubenswrapper[4906]: I1123 07:55:44.931661 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5kqx" event={"ID":"5bbf7a38-6beb-48e6-8723-aa1a70a165a5","Type":"ContainerDied","Data":"0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e"} Nov 23 07:55:44 crc kubenswrapper[4906]: I1123 07:55:44.933575 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5kqx" event={"ID":"5bbf7a38-6beb-48e6-8723-aa1a70a165a5","Type":"ContainerStarted","Data":"d7d93bef2b9af743d0ffabe861c6a32b6023d2150f196e3ff7e9c487eccc7316"} Nov 23 07:55:46 crc kubenswrapper[4906]: I1123 07:55:46.357310 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:55:46 crc kubenswrapper[4906]: E1123 07:55:46.358530 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:55:46 crc kubenswrapper[4906]: I1123 07:55:46.954525 4906 generic.go:334] "Generic (PLEG): container finished" podID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerID="31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211" exitCode=0 Nov 23 07:55:46 crc kubenswrapper[4906]: I1123 07:55:46.954649 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5kqx" event={"ID":"5bbf7a38-6beb-48e6-8723-aa1a70a165a5","Type":"ContainerDied","Data":"31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211"} Nov 23 07:55:47 crc kubenswrapper[4906]: I1123 07:55:47.969174 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5kqx" event={"ID":"5bbf7a38-6beb-48e6-8723-aa1a70a165a5","Type":"ContainerStarted","Data":"fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1"} Nov 23 07:55:47 crc kubenswrapper[4906]: I1123 07:55:47.991910 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5kqx" podStartSLOduration=2.527751388 podStartE2EDuration="4.991889785s" podCreationTimestamp="2025-11-23 07:55:43 +0000 UTC" firstStartedPulling="2025-11-23 07:55:44.933003997 +0000 UTC m=+3960.446395290" lastFinishedPulling="2025-11-23 07:55:47.397142384 +0000 UTC m=+3962.910533687" observedRunningTime="2025-11-23 07:55:47.991444424 +0000 UTC m=+3963.504835717" watchObservedRunningTime="2025-11-23 07:55:47.991889785 +0000 UTC m=+3963.505281088" Nov 23 07:55:53 crc kubenswrapper[4906]: I1123 07:55:53.943631 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:53 crc kubenswrapper[4906]: I1123 07:55:53.946541 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:55:55 crc kubenswrapper[4906]: I1123 07:55:55.008856 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g5kqx" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="registry-server" probeResult="failure" output=< Nov 23 07:55:55 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 07:55:55 crc kubenswrapper[4906]: > Nov 23 07:55:58 crc kubenswrapper[4906]: I1123 07:55:58.357709 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:55:58 crc kubenswrapper[4906]: E1123 07:55:58.358334 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:56:04 crc kubenswrapper[4906]: I1123 07:56:04.018014 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:56:04 crc kubenswrapper[4906]: I1123 07:56:04.085636 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:56:04 crc kubenswrapper[4906]: I1123 07:56:04.264373 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5kqx"] Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.164808 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g5kqx" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="registry-server" containerID="cri-o://fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1" gracePeriod=2 Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.595743 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.691993 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgm6t\" (UniqueName: \"kubernetes.io/projected/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-kube-api-access-kgm6t\") pod \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.692343 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-catalog-content\") pod \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.692484 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-utilities\") pod \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\" (UID: \"5bbf7a38-6beb-48e6-8723-aa1a70a165a5\") " Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.694492 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-utilities" (OuterVolumeSpecName: "utilities") pod "5bbf7a38-6beb-48e6-8723-aa1a70a165a5" (UID: "5bbf7a38-6beb-48e6-8723-aa1a70a165a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.701767 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-kube-api-access-kgm6t" (OuterVolumeSpecName: "kube-api-access-kgm6t") pod "5bbf7a38-6beb-48e6-8723-aa1a70a165a5" (UID: "5bbf7a38-6beb-48e6-8723-aa1a70a165a5"). InnerVolumeSpecName "kube-api-access-kgm6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.789021 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bbf7a38-6beb-48e6-8723-aa1a70a165a5" (UID: "5bbf7a38-6beb-48e6-8723-aa1a70a165a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.794691 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.794735 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgm6t\" (UniqueName: \"kubernetes.io/projected/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-kube-api-access-kgm6t\") on node \"crc\" DevicePath \"\"" Nov 23 07:56:05 crc kubenswrapper[4906]: I1123 07:56:05.794751 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bbf7a38-6beb-48e6-8723-aa1a70a165a5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.177938 4906 generic.go:334] "Generic (PLEG): container finished" podID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerID="fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1" exitCode=0 Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.178005 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5kqx" event={"ID":"5bbf7a38-6beb-48e6-8723-aa1a70a165a5","Type":"ContainerDied","Data":"fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1"} Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.178055 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5kqx" event={"ID":"5bbf7a38-6beb-48e6-8723-aa1a70a165a5","Type":"ContainerDied","Data":"d7d93bef2b9af743d0ffabe861c6a32b6023d2150f196e3ff7e9c487eccc7316"} Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.178080 4906 scope.go:117] "RemoveContainer" containerID="fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.178156 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5kqx" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.228138 4906 scope.go:117] "RemoveContainer" containerID="31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.242509 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5kqx"] Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.253590 4906 scope.go:117] "RemoveContainer" containerID="0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.254618 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g5kqx"] Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.287437 4906 scope.go:117] "RemoveContainer" containerID="fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1" Nov 23 07:56:06 crc kubenswrapper[4906]: E1123 07:56:06.288812 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1\": container with ID starting with fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1 not found: ID does not exist" containerID="fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.288879 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1"} err="failed to get container status \"fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1\": rpc error: code = NotFound desc = could not find container \"fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1\": container with ID starting with fbc6066a39f7e000e33581df7cfce0515a382c57862121f8b62e2920715f34c1 not found: ID does not exist" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.288915 4906 scope.go:117] "RemoveContainer" containerID="31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211" Nov 23 07:56:06 crc kubenswrapper[4906]: E1123 07:56:06.289350 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211\": container with ID starting with 31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211 not found: ID does not exist" containerID="31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.289478 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211"} err="failed to get container status \"31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211\": rpc error: code = NotFound desc = could not find container \"31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211\": container with ID starting with 31678695f631bf8bb4335dcfcdf1b5ea573f43b08872161335eed85c34d31211 not found: ID does not exist" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.289544 4906 scope.go:117] "RemoveContainer" containerID="0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e" Nov 23 07:56:06 crc kubenswrapper[4906]: E1123 07:56:06.290261 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e\": container with ID starting with 0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e not found: ID does not exist" containerID="0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e" Nov 23 07:56:06 crc kubenswrapper[4906]: I1123 07:56:06.290312 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e"} err="failed to get container status \"0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e\": rpc error: code = NotFound desc = could not find container \"0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e\": container with ID starting with 0f861a1580baf6f9afac349d27d7c2af39dcdc791211b96f20f882807abf638e not found: ID does not exist" Nov 23 07:56:07 crc kubenswrapper[4906]: I1123 07:56:07.376028 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" path="/var/lib/kubelet/pods/5bbf7a38-6beb-48e6-8723-aa1a70a165a5/volumes" Nov 23 07:56:10 crc kubenswrapper[4906]: I1123 07:56:10.357116 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:56:10 crc kubenswrapper[4906]: E1123 07:56:10.358456 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:56:24 crc kubenswrapper[4906]: I1123 07:56:24.356597 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:56:24 crc kubenswrapper[4906]: E1123 07:56:24.357934 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:56:36 crc kubenswrapper[4906]: I1123 07:56:36.356744 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:56:36 crc kubenswrapper[4906]: E1123 07:56:36.359327 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:56:51 crc kubenswrapper[4906]: I1123 07:56:51.356767 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:56:51 crc kubenswrapper[4906]: E1123 07:56:51.358295 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:57:02 crc kubenswrapper[4906]: I1123 07:57:02.356563 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:57:02 crc kubenswrapper[4906]: E1123 07:57:02.357796 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:57:14 crc kubenswrapper[4906]: I1123 07:57:14.357373 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:57:14 crc kubenswrapper[4906]: E1123 07:57:14.358343 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:57:28 crc kubenswrapper[4906]: I1123 07:57:28.356465 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:57:28 crc kubenswrapper[4906]: E1123 07:57:28.359177 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.795553 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jkrj6"] Nov 23 07:57:37 crc kubenswrapper[4906]: E1123 07:57:37.797428 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="extract-utilities" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.797486 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="extract-utilities" Nov 23 07:57:37 crc kubenswrapper[4906]: E1123 07:57:37.797501 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="extract-content" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.797509 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="extract-content" Nov 23 07:57:37 crc kubenswrapper[4906]: E1123 07:57:37.797557 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="registry-server" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.797568 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="registry-server" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.797969 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bbf7a38-6beb-48e6-8723-aa1a70a165a5" containerName="registry-server" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.799731 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.842372 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jkrj6"] Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.878469 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-utilities\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.878574 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-catalog-content\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.878635 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcg46\" (UniqueName: \"kubernetes.io/projected/8209516e-f10b-405d-904f-6138392f7e9a-kube-api-access-kcg46\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.979885 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcg46\" (UniqueName: \"kubernetes.io/projected/8209516e-f10b-405d-904f-6138392f7e9a-kube-api-access-kcg46\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.980050 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-utilities\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.980112 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-catalog-content\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.980727 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-utilities\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:37 crc kubenswrapper[4906]: I1123 07:57:37.980884 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-catalog-content\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:38 crc kubenswrapper[4906]: I1123 07:57:38.001832 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcg46\" (UniqueName: \"kubernetes.io/projected/8209516e-f10b-405d-904f-6138392f7e9a-kube-api-access-kcg46\") pod \"certified-operators-jkrj6\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:38 crc kubenswrapper[4906]: I1123 07:57:38.145893 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:38 crc kubenswrapper[4906]: I1123 07:57:38.983930 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jkrj6"] Nov 23 07:57:39 crc kubenswrapper[4906]: I1123 07:57:39.163248 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkrj6" event={"ID":"8209516e-f10b-405d-904f-6138392f7e9a","Type":"ContainerStarted","Data":"9f974f17d19a3c2d702c83e5a7f1be4c51986de447b4593fd9991dda7ec9f3f5"} Nov 23 07:57:39 crc kubenswrapper[4906]: I1123 07:57:39.357330 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:57:39 crc kubenswrapper[4906]: E1123 07:57:39.357805 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:57:40 crc kubenswrapper[4906]: I1123 07:57:40.175537 4906 generic.go:334] "Generic (PLEG): container finished" podID="8209516e-f10b-405d-904f-6138392f7e9a" containerID="827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478" exitCode=0 Nov 23 07:57:40 crc kubenswrapper[4906]: I1123 07:57:40.175620 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkrj6" event={"ID":"8209516e-f10b-405d-904f-6138392f7e9a","Type":"ContainerDied","Data":"827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478"} Nov 23 07:57:41 crc kubenswrapper[4906]: I1123 07:57:41.185759 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkrj6" event={"ID":"8209516e-f10b-405d-904f-6138392f7e9a","Type":"ContainerStarted","Data":"9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f"} Nov 23 07:57:42 crc kubenswrapper[4906]: I1123 07:57:42.201853 4906 generic.go:334] "Generic (PLEG): container finished" podID="8209516e-f10b-405d-904f-6138392f7e9a" containerID="9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f" exitCode=0 Nov 23 07:57:42 crc kubenswrapper[4906]: I1123 07:57:42.201955 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkrj6" event={"ID":"8209516e-f10b-405d-904f-6138392f7e9a","Type":"ContainerDied","Data":"9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f"} Nov 23 07:57:43 crc kubenswrapper[4906]: I1123 07:57:43.216828 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkrj6" event={"ID":"8209516e-f10b-405d-904f-6138392f7e9a","Type":"ContainerStarted","Data":"1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f"} Nov 23 07:57:43 crc kubenswrapper[4906]: I1123 07:57:43.245196 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jkrj6" podStartSLOduration=3.710374694 podStartE2EDuration="6.245171734s" podCreationTimestamp="2025-11-23 07:57:37 +0000 UTC" firstStartedPulling="2025-11-23 07:57:40.177630597 +0000 UTC m=+4075.691021910" lastFinishedPulling="2025-11-23 07:57:42.712427637 +0000 UTC m=+4078.225818950" observedRunningTime="2025-11-23 07:57:43.237543932 +0000 UTC m=+4078.750935235" watchObservedRunningTime="2025-11-23 07:57:43.245171734 +0000 UTC m=+4078.758563037" Nov 23 07:57:48 crc kubenswrapper[4906]: I1123 07:57:48.146574 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:48 crc kubenswrapper[4906]: I1123 07:57:48.147264 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:48 crc kubenswrapper[4906]: I1123 07:57:48.191605 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:48 crc kubenswrapper[4906]: I1123 07:57:48.305584 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:48 crc kubenswrapper[4906]: I1123 07:57:48.425071 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jkrj6"] Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.278523 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jkrj6" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="registry-server" containerID="cri-o://1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f" gracePeriod=2 Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.674746 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.700529 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-utilities\") pod \"8209516e-f10b-405d-904f-6138392f7e9a\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.700627 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-catalog-content\") pod \"8209516e-f10b-405d-904f-6138392f7e9a\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.700807 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcg46\" (UniqueName: \"kubernetes.io/projected/8209516e-f10b-405d-904f-6138392f7e9a-kube-api-access-kcg46\") pod \"8209516e-f10b-405d-904f-6138392f7e9a\" (UID: \"8209516e-f10b-405d-904f-6138392f7e9a\") " Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.702733 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-utilities" (OuterVolumeSpecName: "utilities") pod "8209516e-f10b-405d-904f-6138392f7e9a" (UID: "8209516e-f10b-405d-904f-6138392f7e9a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.706057 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.710934 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8209516e-f10b-405d-904f-6138392f7e9a-kube-api-access-kcg46" (OuterVolumeSpecName: "kube-api-access-kcg46") pod "8209516e-f10b-405d-904f-6138392f7e9a" (UID: "8209516e-f10b-405d-904f-6138392f7e9a"). InnerVolumeSpecName "kube-api-access-kcg46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.759408 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8209516e-f10b-405d-904f-6138392f7e9a" (UID: "8209516e-f10b-405d-904f-6138392f7e9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.808180 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8209516e-f10b-405d-904f-6138392f7e9a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:57:50 crc kubenswrapper[4906]: I1123 07:57:50.808256 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcg46\" (UniqueName: \"kubernetes.io/projected/8209516e-f10b-405d-904f-6138392f7e9a-kube-api-access-kcg46\") on node \"crc\" DevicePath \"\"" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.288159 4906 generic.go:334] "Generic (PLEG): container finished" podID="8209516e-f10b-405d-904f-6138392f7e9a" containerID="1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f" exitCode=0 Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.288218 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkrj6" event={"ID":"8209516e-f10b-405d-904f-6138392f7e9a","Type":"ContainerDied","Data":"1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f"} Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.288772 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkrj6" event={"ID":"8209516e-f10b-405d-904f-6138392f7e9a","Type":"ContainerDied","Data":"9f974f17d19a3c2d702c83e5a7f1be4c51986de447b4593fd9991dda7ec9f3f5"} Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.288808 4906 scope.go:117] "RemoveContainer" containerID="1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.288286 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkrj6" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.310794 4906 scope.go:117] "RemoveContainer" containerID="9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.332197 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jkrj6"] Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.338081 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jkrj6"] Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.348008 4906 scope.go:117] "RemoveContainer" containerID="827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.358069 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:57:51 crc kubenswrapper[4906]: E1123 07:57:51.358518 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.373055 4906 scope.go:117] "RemoveContainer" containerID="1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f" Nov 23 07:57:51 crc kubenswrapper[4906]: E1123 07:57:51.373360 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f\": container with ID starting with 1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f not found: ID does not exist" containerID="1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.373396 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f"} err="failed to get container status \"1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f\": rpc error: code = NotFound desc = could not find container \"1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f\": container with ID starting with 1587e8d1e6c7bb122d634d7e61c9a537b5305b5945f2cf1b4e19e780d705ce0f not found: ID does not exist" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.373422 4906 scope.go:117] "RemoveContainer" containerID="9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f" Nov 23 07:57:51 crc kubenswrapper[4906]: E1123 07:57:51.373638 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f\": container with ID starting with 9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f not found: ID does not exist" containerID="9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.373662 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f"} err="failed to get container status \"9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f\": rpc error: code = NotFound desc = could not find container \"9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f\": container with ID starting with 9f6aef18008935cce6ed8b240909c293fd1a761811cd6eaf7eec78241d04539f not found: ID does not exist" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.373694 4906 scope.go:117] "RemoveContainer" containerID="827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478" Nov 23 07:57:51 crc kubenswrapper[4906]: E1123 07:57:51.373880 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478\": container with ID starting with 827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478 not found: ID does not exist" containerID="827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.373913 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478"} err="failed to get container status \"827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478\": rpc error: code = NotFound desc = could not find container \"827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478\": container with ID starting with 827433ac58c5001b2be57aef60b5ddcfab7aaa50dde17052ebbdf2393887f478 not found: ID does not exist" Nov 23 07:57:51 crc kubenswrapper[4906]: I1123 07:57:51.374844 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8209516e-f10b-405d-904f-6138392f7e9a" path="/var/lib/kubelet/pods/8209516e-f10b-405d-904f-6138392f7e9a/volumes" Nov 23 07:58:02 crc kubenswrapper[4906]: I1123 07:58:02.357441 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:58:02 crc kubenswrapper[4906]: E1123 07:58:02.358888 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:58:17 crc kubenswrapper[4906]: I1123 07:58:17.356919 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:58:17 crc kubenswrapper[4906]: E1123 07:58:17.358077 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:58:30 crc kubenswrapper[4906]: I1123 07:58:30.357813 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:58:30 crc kubenswrapper[4906]: E1123 07:58:30.358953 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:58:45 crc kubenswrapper[4906]: I1123 07:58:45.366333 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:58:45 crc kubenswrapper[4906]: E1123 07:58:45.368002 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 07:58:56 crc kubenswrapper[4906]: I1123 07:58:56.357132 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 07:58:56 crc kubenswrapper[4906]: I1123 07:58:56.979538 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"53a7502a9f0d738e347f40540404b7e171579e228f50124ed57d37b7282cb26c"} Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.178414 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7"] Nov 23 08:00:00 crc kubenswrapper[4906]: E1123 08:00:00.179893 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="extract-utilities" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.180076 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="extract-utilities" Nov 23 08:00:00 crc kubenswrapper[4906]: E1123 08:00:00.180098 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="extract-content" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.180112 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="extract-content" Nov 23 08:00:00 crc kubenswrapper[4906]: E1123 08:00:00.180132 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="registry-server" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.180146 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="registry-server" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.180427 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="8209516e-f10b-405d-904f-6138392f7e9a" containerName="registry-server" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.181333 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.187973 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.190941 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.220311 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7"] Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.288662 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d28ea74-f343-4b45-a4a0-56dad68591db-config-volume\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.288754 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zfvk\" (UniqueName: \"kubernetes.io/projected/1d28ea74-f343-4b45-a4a0-56dad68591db-kube-api-access-4zfvk\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.288781 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d28ea74-f343-4b45-a4a0-56dad68591db-secret-volume\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.389989 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d28ea74-f343-4b45-a4a0-56dad68591db-config-volume\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.390064 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zfvk\" (UniqueName: \"kubernetes.io/projected/1d28ea74-f343-4b45-a4a0-56dad68591db-kube-api-access-4zfvk\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.390090 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d28ea74-f343-4b45-a4a0-56dad68591db-secret-volume\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.391140 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d28ea74-f343-4b45-a4a0-56dad68591db-config-volume\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.398343 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d28ea74-f343-4b45-a4a0-56dad68591db-secret-volume\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.414326 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zfvk\" (UniqueName: \"kubernetes.io/projected/1d28ea74-f343-4b45-a4a0-56dad68591db-kube-api-access-4zfvk\") pod \"collect-profiles-29398080-stbg7\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.513624 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:00 crc kubenswrapper[4906]: I1123 08:00:00.817529 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7"] Nov 23 08:00:00 crc kubenswrapper[4906]: W1123 08:00:00.824615 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d28ea74_f343_4b45_a4a0_56dad68591db.slice/crio-f8ec9c08863d3e58d4071dd4a5ceb7cbd5ce25cfca870ec0c56f21c2327e04d5 WatchSource:0}: Error finding container f8ec9c08863d3e58d4071dd4a5ceb7cbd5ce25cfca870ec0c56f21c2327e04d5: Status 404 returned error can't find the container with id f8ec9c08863d3e58d4071dd4a5ceb7cbd5ce25cfca870ec0c56f21c2327e04d5 Nov 23 08:00:01 crc kubenswrapper[4906]: I1123 08:00:01.160455 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" event={"ID":"1d28ea74-f343-4b45-a4a0-56dad68591db","Type":"ContainerStarted","Data":"3942801e1ce014b645ae0f5047f24ee84d2b3da04d1b40a9792d80de47a708a7"} Nov 23 08:00:01 crc kubenswrapper[4906]: I1123 08:00:01.160534 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" event={"ID":"1d28ea74-f343-4b45-a4a0-56dad68591db","Type":"ContainerStarted","Data":"f8ec9c08863d3e58d4071dd4a5ceb7cbd5ce25cfca870ec0c56f21c2327e04d5"} Nov 23 08:00:01 crc kubenswrapper[4906]: I1123 08:00:01.188021 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" podStartSLOduration=1.18799676 podStartE2EDuration="1.18799676s" podCreationTimestamp="2025-11-23 08:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 08:00:01.180610395 +0000 UTC m=+4216.694001698" watchObservedRunningTime="2025-11-23 08:00:01.18799676 +0000 UTC m=+4216.701388073" Nov 23 08:00:02 crc kubenswrapper[4906]: I1123 08:00:02.173736 4906 generic.go:334] "Generic (PLEG): container finished" podID="1d28ea74-f343-4b45-a4a0-56dad68591db" containerID="3942801e1ce014b645ae0f5047f24ee84d2b3da04d1b40a9792d80de47a708a7" exitCode=0 Nov 23 08:00:02 crc kubenswrapper[4906]: I1123 08:00:02.174148 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" event={"ID":"1d28ea74-f343-4b45-a4a0-56dad68591db","Type":"ContainerDied","Data":"3942801e1ce014b645ae0f5047f24ee84d2b3da04d1b40a9792d80de47a708a7"} Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.544993 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.646627 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d28ea74-f343-4b45-a4a0-56dad68591db-secret-volume\") pod \"1d28ea74-f343-4b45-a4a0-56dad68591db\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.646804 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d28ea74-f343-4b45-a4a0-56dad68591db-config-volume\") pod \"1d28ea74-f343-4b45-a4a0-56dad68591db\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.646865 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zfvk\" (UniqueName: \"kubernetes.io/projected/1d28ea74-f343-4b45-a4a0-56dad68591db-kube-api-access-4zfvk\") pod \"1d28ea74-f343-4b45-a4a0-56dad68591db\" (UID: \"1d28ea74-f343-4b45-a4a0-56dad68591db\") " Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.647624 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d28ea74-f343-4b45-a4a0-56dad68591db-config-volume" (OuterVolumeSpecName: "config-volume") pod "1d28ea74-f343-4b45-a4a0-56dad68591db" (UID: "1d28ea74-f343-4b45-a4a0-56dad68591db"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.656448 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d28ea74-f343-4b45-a4a0-56dad68591db-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1d28ea74-f343-4b45-a4a0-56dad68591db" (UID: "1d28ea74-f343-4b45-a4a0-56dad68591db"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.660057 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d28ea74-f343-4b45-a4a0-56dad68591db-kube-api-access-4zfvk" (OuterVolumeSpecName: "kube-api-access-4zfvk") pod "1d28ea74-f343-4b45-a4a0-56dad68591db" (UID: "1d28ea74-f343-4b45-a4a0-56dad68591db"). InnerVolumeSpecName "kube-api-access-4zfvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.749080 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d28ea74-f343-4b45-a4a0-56dad68591db-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.749136 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d28ea74-f343-4b45-a4a0-56dad68591db-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:00:03 crc kubenswrapper[4906]: I1123 08:00:03.749151 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zfvk\" (UniqueName: \"kubernetes.io/projected/1d28ea74-f343-4b45-a4a0-56dad68591db-kube-api-access-4zfvk\") on node \"crc\" DevicePath \"\"" Nov 23 08:00:04 crc kubenswrapper[4906]: I1123 08:00:04.192202 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" event={"ID":"1d28ea74-f343-4b45-a4a0-56dad68591db","Type":"ContainerDied","Data":"f8ec9c08863d3e58d4071dd4a5ceb7cbd5ce25cfca870ec0c56f21c2327e04d5"} Nov 23 08:00:04 crc kubenswrapper[4906]: I1123 08:00:04.192758 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ec9c08863d3e58d4071dd4a5ceb7cbd5ce25cfca870ec0c56f21c2327e04d5" Nov 23 08:00:04 crc kubenswrapper[4906]: I1123 08:00:04.192339 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7" Nov 23 08:00:04 crc kubenswrapper[4906]: I1123 08:00:04.283090 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj"] Nov 23 08:00:04 crc kubenswrapper[4906]: I1123 08:00:04.289244 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-hh8jj"] Nov 23 08:00:05 crc kubenswrapper[4906]: I1123 08:00:05.369433 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fcb8821-fed8-4298-97c0-458ef9c6f6c2" path="/var/lib/kubelet/pods/6fcb8821-fed8-4298-97c0-458ef9c6f6c2/volumes" Nov 23 08:00:57 crc kubenswrapper[4906]: I1123 08:00:57.660405 4906 scope.go:117] "RemoveContainer" containerID="2bec6b3863851e923a4cbe255526d460a7ce79bb7b309ab01a6cb0cb10cc6cac" Nov 23 08:00:57 crc kubenswrapper[4906]: I1123 08:00:57.697490 4906 scope.go:117] "RemoveContainer" containerID="41a8bcdd6ef8d6168db951a4515f0e58d0247a4dae56d6ca0d85cd3d6c9d30b2" Nov 23 08:00:57 crc kubenswrapper[4906]: I1123 08:00:57.742379 4906 scope.go:117] "RemoveContainer" containerID="1e02e4838a754b86154bf6212219c7d23c6cd07ca317e019d6ca3cb47876cdaf" Nov 23 08:00:57 crc kubenswrapper[4906]: I1123 08:00:57.773603 4906 scope.go:117] "RemoveContainer" containerID="a755255fff7ae443ab7c7bdfe587c3014f547385266101a3bb3e629582180cb7" Nov 23 08:01:20 crc kubenswrapper[4906]: I1123 08:01:20.946358 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:01:20 crc kubenswrapper[4906]: I1123 08:01:20.947151 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:01:50 crc kubenswrapper[4906]: I1123 08:01:50.946199 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:01:50 crc kubenswrapper[4906]: I1123 08:01:50.947303 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:02:20 crc kubenswrapper[4906]: I1123 08:02:20.945601 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:02:20 crc kubenswrapper[4906]: I1123 08:02:20.946548 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:02:20 crc kubenswrapper[4906]: I1123 08:02:20.946618 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:02:20 crc kubenswrapper[4906]: I1123 08:02:20.947627 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53a7502a9f0d738e347f40540404b7e171579e228f50124ed57d37b7282cb26c"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:02:20 crc kubenswrapper[4906]: I1123 08:02:20.947730 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://53a7502a9f0d738e347f40540404b7e171579e228f50124ed57d37b7282cb26c" gracePeriod=600 Nov 23 08:02:21 crc kubenswrapper[4906]: I1123 08:02:21.634667 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="53a7502a9f0d738e347f40540404b7e171579e228f50124ed57d37b7282cb26c" exitCode=0 Nov 23 08:02:21 crc kubenswrapper[4906]: I1123 08:02:21.634751 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"53a7502a9f0d738e347f40540404b7e171579e228f50124ed57d37b7282cb26c"} Nov 23 08:02:21 crc kubenswrapper[4906]: I1123 08:02:21.635150 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45"} Nov 23 08:02:21 crc kubenswrapper[4906]: I1123 08:02:21.635187 4906 scope.go:117] "RemoveContainer" containerID="030b7deb6eca38f3fb579b492a4b768424cf9202a3f6e79742ed51d7b25d1b51" Nov 23 08:03:03 crc kubenswrapper[4906]: I1123 08:03:02.862846 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-zhqrw" podUID="8a5e4a09-9b16-435b-bc05-00d6d174ede3" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.200669 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b942d"] Nov 23 08:04:50 crc kubenswrapper[4906]: E1123 08:04:50.202324 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d28ea74-f343-4b45-a4a0-56dad68591db" containerName="collect-profiles" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.202352 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d28ea74-f343-4b45-a4a0-56dad68591db" containerName="collect-profiles" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.202644 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d28ea74-f343-4b45-a4a0-56dad68591db" containerName="collect-profiles" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.204677 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.230919 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b942d"] Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.384900 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l69lm\" (UniqueName: \"kubernetes.io/projected/a1373ae5-8cee-4aa2-9032-9824e36efad0-kube-api-access-l69lm\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.385051 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-utilities\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.385358 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-catalog-content\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.486922 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-catalog-content\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.486997 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l69lm\" (UniqueName: \"kubernetes.io/projected/a1373ae5-8cee-4aa2-9032-9824e36efad0-kube-api-access-l69lm\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.487055 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-utilities\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.487784 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-utilities\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.487830 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-catalog-content\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.514657 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l69lm\" (UniqueName: \"kubernetes.io/projected/a1373ae5-8cee-4aa2-9032-9824e36efad0-kube-api-access-l69lm\") pod \"community-operators-b942d\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.536649 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.824114 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b942d"] Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.945531 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:04:50 crc kubenswrapper[4906]: I1123 08:04:50.946111 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:04:51 crc kubenswrapper[4906]: I1123 08:04:51.233724 4906 generic.go:334] "Generic (PLEG): container finished" podID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerID="4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82" exitCode=0 Nov 23 08:04:51 crc kubenswrapper[4906]: I1123 08:04:51.233846 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b942d" event={"ID":"a1373ae5-8cee-4aa2-9032-9824e36efad0","Type":"ContainerDied","Data":"4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82"} Nov 23 08:04:51 crc kubenswrapper[4906]: I1123 08:04:51.233920 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b942d" event={"ID":"a1373ae5-8cee-4aa2-9032-9824e36efad0","Type":"ContainerStarted","Data":"1b26cfa63ae10425c0a43dce3cae1533e06e726e28d748a97dd6fc40f91103d0"} Nov 23 08:04:51 crc kubenswrapper[4906]: I1123 08:04:51.236407 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:04:52 crc kubenswrapper[4906]: I1123 08:04:52.244365 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b942d" event={"ID":"a1373ae5-8cee-4aa2-9032-9824e36efad0","Type":"ContainerStarted","Data":"662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c"} Nov 23 08:04:53 crc kubenswrapper[4906]: I1123 08:04:53.257296 4906 generic.go:334] "Generic (PLEG): container finished" podID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerID="662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c" exitCode=0 Nov 23 08:04:53 crc kubenswrapper[4906]: I1123 08:04:53.257390 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b942d" event={"ID":"a1373ae5-8cee-4aa2-9032-9824e36efad0","Type":"ContainerDied","Data":"662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c"} Nov 23 08:04:54 crc kubenswrapper[4906]: I1123 08:04:54.288661 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b942d" event={"ID":"a1373ae5-8cee-4aa2-9032-9824e36efad0","Type":"ContainerStarted","Data":"1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6"} Nov 23 08:04:54 crc kubenswrapper[4906]: I1123 08:04:54.313967 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b942d" podStartSLOduration=1.895794481 podStartE2EDuration="4.313941373s" podCreationTimestamp="2025-11-23 08:04:50 +0000 UTC" firstStartedPulling="2025-11-23 08:04:51.23591787 +0000 UTC m=+4506.749309173" lastFinishedPulling="2025-11-23 08:04:53.654064722 +0000 UTC m=+4509.167456065" observedRunningTime="2025-11-23 08:04:54.307553725 +0000 UTC m=+4509.820945038" watchObservedRunningTime="2025-11-23 08:04:54.313941373 +0000 UTC m=+4509.827332686" Nov 23 08:05:00 crc kubenswrapper[4906]: I1123 08:05:00.537744 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:05:00 crc kubenswrapper[4906]: I1123 08:05:00.538603 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:05:00 crc kubenswrapper[4906]: I1123 08:05:00.618313 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:05:01 crc kubenswrapper[4906]: I1123 08:05:01.421313 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:05:01 crc kubenswrapper[4906]: I1123 08:05:01.490092 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b942d"] Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.389073 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b942d" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="registry-server" containerID="cri-o://1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6" gracePeriod=2 Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.857822 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.923513 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-catalog-content\") pod \"a1373ae5-8cee-4aa2-9032-9824e36efad0\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.923632 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l69lm\" (UniqueName: \"kubernetes.io/projected/a1373ae5-8cee-4aa2-9032-9824e36efad0-kube-api-access-l69lm\") pod \"a1373ae5-8cee-4aa2-9032-9824e36efad0\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.923772 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-utilities\") pod \"a1373ae5-8cee-4aa2-9032-9824e36efad0\" (UID: \"a1373ae5-8cee-4aa2-9032-9824e36efad0\") " Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.924656 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-utilities" (OuterVolumeSpecName: "utilities") pod "a1373ae5-8cee-4aa2-9032-9824e36efad0" (UID: "a1373ae5-8cee-4aa2-9032-9824e36efad0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.931392 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1373ae5-8cee-4aa2-9032-9824e36efad0-kube-api-access-l69lm" (OuterVolumeSpecName: "kube-api-access-l69lm") pod "a1373ae5-8cee-4aa2-9032-9824e36efad0" (UID: "a1373ae5-8cee-4aa2-9032-9824e36efad0"). InnerVolumeSpecName "kube-api-access-l69lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:05:03 crc kubenswrapper[4906]: I1123 08:05:03.984630 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1373ae5-8cee-4aa2-9032-9824e36efad0" (UID: "a1373ae5-8cee-4aa2-9032-9824e36efad0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.026305 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.026354 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l69lm\" (UniqueName: \"kubernetes.io/projected/a1373ae5-8cee-4aa2-9032-9824e36efad0-kube-api-access-l69lm\") on node \"crc\" DevicePath \"\"" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.026372 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1373ae5-8cee-4aa2-9032-9824e36efad0-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.411404 4906 generic.go:334] "Generic (PLEG): container finished" podID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerID="1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6" exitCode=0 Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.411498 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b942d" event={"ID":"a1373ae5-8cee-4aa2-9032-9824e36efad0","Type":"ContainerDied","Data":"1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6"} Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.411538 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b942d" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.411582 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b942d" event={"ID":"a1373ae5-8cee-4aa2-9032-9824e36efad0","Type":"ContainerDied","Data":"1b26cfa63ae10425c0a43dce3cae1533e06e726e28d748a97dd6fc40f91103d0"} Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.411636 4906 scope.go:117] "RemoveContainer" containerID="1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.462219 4906 scope.go:117] "RemoveContainer" containerID="662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.488581 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b942d"] Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.499479 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b942d"] Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.514520 4906 scope.go:117] "RemoveContainer" containerID="4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.548914 4906 scope.go:117] "RemoveContainer" containerID="1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6" Nov 23 08:05:04 crc kubenswrapper[4906]: E1123 08:05:04.549524 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6\": container with ID starting with 1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6 not found: ID does not exist" containerID="1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.549569 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6"} err="failed to get container status \"1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6\": rpc error: code = NotFound desc = could not find container \"1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6\": container with ID starting with 1f849d6d62c73b014fc14a9093956937c2b631ca09cccece0826533a1379f1d6 not found: ID does not exist" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.549612 4906 scope.go:117] "RemoveContainer" containerID="662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c" Nov 23 08:05:04 crc kubenswrapper[4906]: E1123 08:05:04.550053 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c\": container with ID starting with 662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c not found: ID does not exist" containerID="662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.550130 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c"} err="failed to get container status \"662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c\": rpc error: code = NotFound desc = could not find container \"662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c\": container with ID starting with 662fdc3f1d0cafed54b57a0904573ad712ea6dbedc0dd542ed6a5c0c0787616c not found: ID does not exist" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.550182 4906 scope.go:117] "RemoveContainer" containerID="4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82" Nov 23 08:05:04 crc kubenswrapper[4906]: E1123 08:05:04.550997 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82\": container with ID starting with 4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82 not found: ID does not exist" containerID="4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82" Nov 23 08:05:04 crc kubenswrapper[4906]: I1123 08:05:04.551034 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82"} err="failed to get container status \"4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82\": rpc error: code = NotFound desc = could not find container \"4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82\": container with ID starting with 4444e274bc913f0b1a14ae826efdb9919b801a8832430220776a1af9c1d59b82 not found: ID does not exist" Nov 23 08:05:04 crc kubenswrapper[4906]: E1123 08:05:04.679127 4906 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1373ae5_8cee_4aa2_9032_9824e36efad0.slice/crio-1b26cfa63ae10425c0a43dce3cae1533e06e726e28d748a97dd6fc40f91103d0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1373ae5_8cee_4aa2_9032_9824e36efad0.slice\": RecentStats: unable to find data in memory cache]" Nov 23 08:05:05 crc kubenswrapper[4906]: I1123 08:05:05.377601 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" path="/var/lib/kubelet/pods/a1373ae5-8cee-4aa2-9032-9824e36efad0/volumes" Nov 23 08:05:20 crc kubenswrapper[4906]: I1123 08:05:20.946591 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:05:20 crc kubenswrapper[4906]: I1123 08:05:20.947482 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:05:50 crc kubenswrapper[4906]: I1123 08:05:50.946230 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:05:50 crc kubenswrapper[4906]: I1123 08:05:50.947011 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:05:50 crc kubenswrapper[4906]: I1123 08:05:50.947090 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:05:50 crc kubenswrapper[4906]: I1123 08:05:50.948136 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:05:50 crc kubenswrapper[4906]: I1123 08:05:50.948239 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" gracePeriod=600 Nov 23 08:05:51 crc kubenswrapper[4906]: E1123 08:05:51.082752 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:05:51 crc kubenswrapper[4906]: I1123 08:05:51.882036 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" exitCode=0 Nov 23 08:05:51 crc kubenswrapper[4906]: I1123 08:05:51.882128 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45"} Nov 23 08:05:51 crc kubenswrapper[4906]: I1123 08:05:51.882230 4906 scope.go:117] "RemoveContainer" containerID="53a7502a9f0d738e347f40540404b7e171579e228f50124ed57d37b7282cb26c" Nov 23 08:05:51 crc kubenswrapper[4906]: I1123 08:05:51.883148 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:05:51 crc kubenswrapper[4906]: E1123 08:05:51.883606 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:06:07 crc kubenswrapper[4906]: I1123 08:06:07.357361 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:06:07 crc kubenswrapper[4906]: E1123 08:06:07.359542 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.426257 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w62pr"] Nov 23 08:06:15 crc kubenswrapper[4906]: E1123 08:06:15.427196 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="extract-utilities" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.427214 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="extract-utilities" Nov 23 08:06:15 crc kubenswrapper[4906]: E1123 08:06:15.427275 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="registry-server" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.427282 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="registry-server" Nov 23 08:06:15 crc kubenswrapper[4906]: E1123 08:06:15.427292 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="extract-content" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.427298 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="extract-content" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.427519 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1373ae5-8cee-4aa2-9032-9824e36efad0" containerName="registry-server" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.429110 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w62pr"] Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.429224 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.522065 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-catalog-content\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.522123 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wccdf\" (UniqueName: \"kubernetes.io/projected/9475af7c-5408-48fc-a108-eeea194ee490-kube-api-access-wccdf\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.522151 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-utilities\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.623633 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-catalog-content\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.623726 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wccdf\" (UniqueName: \"kubernetes.io/projected/9475af7c-5408-48fc-a108-eeea194ee490-kube-api-access-wccdf\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.623759 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-utilities\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.624568 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-utilities\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.624707 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-catalog-content\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.654199 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wccdf\" (UniqueName: \"kubernetes.io/projected/9475af7c-5408-48fc-a108-eeea194ee490-kube-api-access-wccdf\") pod \"redhat-operators-w62pr\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:15 crc kubenswrapper[4906]: I1123 08:06:15.756116 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:16 crc kubenswrapper[4906]: I1123 08:06:16.261082 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w62pr"] Nov 23 08:06:17 crc kubenswrapper[4906]: I1123 08:06:17.157278 4906 generic.go:334] "Generic (PLEG): container finished" podID="9475af7c-5408-48fc-a108-eeea194ee490" containerID="892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243" exitCode=0 Nov 23 08:06:17 crc kubenswrapper[4906]: I1123 08:06:17.157353 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w62pr" event={"ID":"9475af7c-5408-48fc-a108-eeea194ee490","Type":"ContainerDied","Data":"892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243"} Nov 23 08:06:17 crc kubenswrapper[4906]: I1123 08:06:17.158101 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w62pr" event={"ID":"9475af7c-5408-48fc-a108-eeea194ee490","Type":"ContainerStarted","Data":"8905fcd0d5d398cee2df81f721a205361ecce4575c330059c2f9e2c2bf8a5e6e"} Nov 23 08:06:18 crc kubenswrapper[4906]: I1123 08:06:18.169890 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w62pr" event={"ID":"9475af7c-5408-48fc-a108-eeea194ee490","Type":"ContainerStarted","Data":"bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8"} Nov 23 08:06:19 crc kubenswrapper[4906]: I1123 08:06:19.182006 4906 generic.go:334] "Generic (PLEG): container finished" podID="9475af7c-5408-48fc-a108-eeea194ee490" containerID="bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8" exitCode=0 Nov 23 08:06:19 crc kubenswrapper[4906]: I1123 08:06:19.182154 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w62pr" event={"ID":"9475af7c-5408-48fc-a108-eeea194ee490","Type":"ContainerDied","Data":"bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8"} Nov 23 08:06:20 crc kubenswrapper[4906]: I1123 08:06:20.197091 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w62pr" event={"ID":"9475af7c-5408-48fc-a108-eeea194ee490","Type":"ContainerStarted","Data":"b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8"} Nov 23 08:06:20 crc kubenswrapper[4906]: I1123 08:06:20.225980 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w62pr" podStartSLOduration=2.737301775 podStartE2EDuration="5.225951969s" podCreationTimestamp="2025-11-23 08:06:15 +0000 UTC" firstStartedPulling="2025-11-23 08:06:17.160980738 +0000 UTC m=+4592.674372071" lastFinishedPulling="2025-11-23 08:06:19.649630932 +0000 UTC m=+4595.163022265" observedRunningTime="2025-11-23 08:06:20.218811321 +0000 UTC m=+4595.732202684" watchObservedRunningTime="2025-11-23 08:06:20.225951969 +0000 UTC m=+4595.739343272" Nov 23 08:06:20 crc kubenswrapper[4906]: I1123 08:06:20.356724 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:06:20 crc kubenswrapper[4906]: E1123 08:06:20.357105 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:06:25 crc kubenswrapper[4906]: I1123 08:06:25.756404 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:25 crc kubenswrapper[4906]: I1123 08:06:25.757597 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:26 crc kubenswrapper[4906]: I1123 08:06:26.804030 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w62pr" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="registry-server" probeResult="failure" output=< Nov 23 08:06:26 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 08:06:26 crc kubenswrapper[4906]: > Nov 23 08:06:33 crc kubenswrapper[4906]: I1123 08:06:33.357179 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:06:33 crc kubenswrapper[4906]: E1123 08:06:33.358342 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:06:35 crc kubenswrapper[4906]: I1123 08:06:35.826479 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:35 crc kubenswrapper[4906]: I1123 08:06:35.882622 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:36 crc kubenswrapper[4906]: I1123 08:06:36.078420 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w62pr"] Nov 23 08:06:37 crc kubenswrapper[4906]: I1123 08:06:37.402918 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w62pr" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="registry-server" containerID="cri-o://b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8" gracePeriod=2 Nov 23 08:06:37 crc kubenswrapper[4906]: I1123 08:06:37.848618 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.020716 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-catalog-content\") pod \"9475af7c-5408-48fc-a108-eeea194ee490\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.020865 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-utilities\") pod \"9475af7c-5408-48fc-a108-eeea194ee490\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.020979 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wccdf\" (UniqueName: \"kubernetes.io/projected/9475af7c-5408-48fc-a108-eeea194ee490-kube-api-access-wccdf\") pod \"9475af7c-5408-48fc-a108-eeea194ee490\" (UID: \"9475af7c-5408-48fc-a108-eeea194ee490\") " Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.022174 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-utilities" (OuterVolumeSpecName: "utilities") pod "9475af7c-5408-48fc-a108-eeea194ee490" (UID: "9475af7c-5408-48fc-a108-eeea194ee490"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.027283 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9475af7c-5408-48fc-a108-eeea194ee490-kube-api-access-wccdf" (OuterVolumeSpecName: "kube-api-access-wccdf") pod "9475af7c-5408-48fc-a108-eeea194ee490" (UID: "9475af7c-5408-48fc-a108-eeea194ee490"). InnerVolumeSpecName "kube-api-access-wccdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.114203 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9475af7c-5408-48fc-a108-eeea194ee490" (UID: "9475af7c-5408-48fc-a108-eeea194ee490"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.122432 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.122466 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wccdf\" (UniqueName: \"kubernetes.io/projected/9475af7c-5408-48fc-a108-eeea194ee490-kube-api-access-wccdf\") on node \"crc\" DevicePath \"\"" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.122476 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9475af7c-5408-48fc-a108-eeea194ee490-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.416147 4906 generic.go:334] "Generic (PLEG): container finished" podID="9475af7c-5408-48fc-a108-eeea194ee490" containerID="b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8" exitCode=0 Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.416215 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w62pr" event={"ID":"9475af7c-5408-48fc-a108-eeea194ee490","Type":"ContainerDied","Data":"b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8"} Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.416252 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w62pr" event={"ID":"9475af7c-5408-48fc-a108-eeea194ee490","Type":"ContainerDied","Data":"8905fcd0d5d398cee2df81f721a205361ecce4575c330059c2f9e2c2bf8a5e6e"} Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.416275 4906 scope.go:117] "RemoveContainer" containerID="b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.416273 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w62pr" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.445150 4906 scope.go:117] "RemoveContainer" containerID="bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.454450 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w62pr"] Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.462315 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w62pr"] Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.487752 4906 scope.go:117] "RemoveContainer" containerID="892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.511597 4906 scope.go:117] "RemoveContainer" containerID="b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8" Nov 23 08:06:38 crc kubenswrapper[4906]: E1123 08:06:38.512093 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8\": container with ID starting with b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8 not found: ID does not exist" containerID="b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.512152 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8"} err="failed to get container status \"b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8\": rpc error: code = NotFound desc = could not find container \"b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8\": container with ID starting with b958239d99d4b45c928bd047e197245c014f999c7c59becae17410d53f15dbf8 not found: ID does not exist" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.512190 4906 scope.go:117] "RemoveContainer" containerID="bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8" Nov 23 08:06:38 crc kubenswrapper[4906]: E1123 08:06:38.512612 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8\": container with ID starting with bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8 not found: ID does not exist" containerID="bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.512703 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8"} err="failed to get container status \"bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8\": rpc error: code = NotFound desc = could not find container \"bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8\": container with ID starting with bb24c73c013845943a8243b691bd0eebaa58fab03baf883ecc3d6643a1a74cb8 not found: ID does not exist" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.512758 4906 scope.go:117] "RemoveContainer" containerID="892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243" Nov 23 08:06:38 crc kubenswrapper[4906]: E1123 08:06:38.513127 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243\": container with ID starting with 892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243 not found: ID does not exist" containerID="892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243" Nov 23 08:06:38 crc kubenswrapper[4906]: I1123 08:06:38.513172 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243"} err="failed to get container status \"892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243\": rpc error: code = NotFound desc = could not find container \"892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243\": container with ID starting with 892b2d584291853eaad3baf70279be634dd800688bfb7af102134ed7afb16243 not found: ID does not exist" Nov 23 08:06:39 crc kubenswrapper[4906]: I1123 08:06:39.369806 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9475af7c-5408-48fc-a108-eeea194ee490" path="/var/lib/kubelet/pods/9475af7c-5408-48fc-a108-eeea194ee490/volumes" Nov 23 08:06:47 crc kubenswrapper[4906]: I1123 08:06:47.358393 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:06:47 crc kubenswrapper[4906]: E1123 08:06:47.359776 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:07:00 crc kubenswrapper[4906]: I1123 08:07:00.358049 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:07:00 crc kubenswrapper[4906]: E1123 08:07:00.359468 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:07:13 crc kubenswrapper[4906]: I1123 08:07:13.356904 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:07:13 crc kubenswrapper[4906]: E1123 08:07:13.357901 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:07:24 crc kubenswrapper[4906]: I1123 08:07:24.357548 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:07:24 crc kubenswrapper[4906]: E1123 08:07:24.358567 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:07:35 crc kubenswrapper[4906]: I1123 08:07:35.361083 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:07:35 crc kubenswrapper[4906]: E1123 08:07:35.361961 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:07:49 crc kubenswrapper[4906]: I1123 08:07:49.357489 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:07:49 crc kubenswrapper[4906]: E1123 08:07:49.358865 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:08:02 crc kubenswrapper[4906]: I1123 08:08:02.357775 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:08:02 crc kubenswrapper[4906]: E1123 08:08:02.359058 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:08:16 crc kubenswrapper[4906]: I1123 08:08:16.356643 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:08:16 crc kubenswrapper[4906]: E1123 08:08:16.358208 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.474935 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v4hh7"] Nov 23 08:08:19 crc kubenswrapper[4906]: E1123 08:08:19.476116 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="extract-content" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.476133 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="extract-content" Nov 23 08:08:19 crc kubenswrapper[4906]: E1123 08:08:19.476149 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="extract-utilities" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.476156 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="extract-utilities" Nov 23 08:08:19 crc kubenswrapper[4906]: E1123 08:08:19.476165 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="registry-server" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.476172 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="registry-server" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.476338 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="9475af7c-5408-48fc-a108-eeea194ee490" containerName="registry-server" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.477591 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.513878 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4hh7"] Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.573631 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-utilities\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.573769 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-catalog-content\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.573805 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/09b96406-11c2-49bc-a54d-4d68af3cc9ef-kube-api-access-85z98\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.674999 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-catalog-content\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.675067 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/09b96406-11c2-49bc-a54d-4d68af3cc9ef-kube-api-access-85z98\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.675163 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-utilities\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.675761 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-catalog-content\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.675957 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-utilities\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.701382 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/09b96406-11c2-49bc-a54d-4d68af3cc9ef-kube-api-access-85z98\") pod \"certified-operators-v4hh7\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:19 crc kubenswrapper[4906]: I1123 08:08:19.802958 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:20 crc kubenswrapper[4906]: I1123 08:08:20.277376 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4hh7"] Nov 23 08:08:20 crc kubenswrapper[4906]: I1123 08:08:20.471038 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerStarted","Data":"908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8"} Nov 23 08:08:20 crc kubenswrapper[4906]: I1123 08:08:20.471098 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerStarted","Data":"9ba744da3451a279c9364ebbb8f22bb5aa42e0b33a222896d4cf67821c068dd4"} Nov 23 08:08:21 crc kubenswrapper[4906]: I1123 08:08:21.485031 4906 generic.go:334] "Generic (PLEG): container finished" podID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerID="908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8" exitCode=0 Nov 23 08:08:21 crc kubenswrapper[4906]: I1123 08:08:21.485105 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerDied","Data":"908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8"} Nov 23 08:08:22 crc kubenswrapper[4906]: I1123 08:08:22.533698 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerStarted","Data":"677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012"} Nov 23 08:08:23 crc kubenswrapper[4906]: I1123 08:08:23.549840 4906 generic.go:334] "Generic (PLEG): container finished" podID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerID="677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012" exitCode=0 Nov 23 08:08:23 crc kubenswrapper[4906]: I1123 08:08:23.549908 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerDied","Data":"677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012"} Nov 23 08:08:24 crc kubenswrapper[4906]: I1123 08:08:24.562207 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerStarted","Data":"9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371"} Nov 23 08:08:24 crc kubenswrapper[4906]: I1123 08:08:24.592075 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v4hh7" podStartSLOduration=3.08867725 podStartE2EDuration="5.592050517s" podCreationTimestamp="2025-11-23 08:08:19 +0000 UTC" firstStartedPulling="2025-11-23 08:08:21.48872886 +0000 UTC m=+4717.002120203" lastFinishedPulling="2025-11-23 08:08:23.992102167 +0000 UTC m=+4719.505493470" observedRunningTime="2025-11-23 08:08:24.589830199 +0000 UTC m=+4720.103221512" watchObservedRunningTime="2025-11-23 08:08:24.592050517 +0000 UTC m=+4720.105441830" Nov 23 08:08:29 crc kubenswrapper[4906]: I1123 08:08:29.803299 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:29 crc kubenswrapper[4906]: I1123 08:08:29.804114 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:29 crc kubenswrapper[4906]: I1123 08:08:29.882571 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:30 crc kubenswrapper[4906]: I1123 08:08:30.357095 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:08:30 crc kubenswrapper[4906]: E1123 08:08:30.357604 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:08:30 crc kubenswrapper[4906]: I1123 08:08:30.693446 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:30 crc kubenswrapper[4906]: I1123 08:08:30.772825 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4hh7"] Nov 23 08:08:32 crc kubenswrapper[4906]: I1123 08:08:32.639737 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v4hh7" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="registry-server" containerID="cri-o://9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371" gracePeriod=2 Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.138446 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.328713 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-utilities\") pod \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.328855 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-catalog-content\") pod \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.328921 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/09b96406-11c2-49bc-a54d-4d68af3cc9ef-kube-api-access-85z98\") pod \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\" (UID: \"09b96406-11c2-49bc-a54d-4d68af3cc9ef\") " Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.331350 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-utilities" (OuterVolumeSpecName: "utilities") pod "09b96406-11c2-49bc-a54d-4d68af3cc9ef" (UID: "09b96406-11c2-49bc-a54d-4d68af3cc9ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.340207 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b96406-11c2-49bc-a54d-4d68af3cc9ef-kube-api-access-85z98" (OuterVolumeSpecName: "kube-api-access-85z98") pod "09b96406-11c2-49bc-a54d-4d68af3cc9ef" (UID: "09b96406-11c2-49bc-a54d-4d68af3cc9ef"). InnerVolumeSpecName "kube-api-access-85z98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.381740 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09b96406-11c2-49bc-a54d-4d68af3cc9ef" (UID: "09b96406-11c2-49bc-a54d-4d68af3cc9ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.431071 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.431327 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b96406-11c2-49bc-a54d-4d68af3cc9ef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.431401 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/09b96406-11c2-49bc-a54d-4d68af3cc9ef-kube-api-access-85z98\") on node \"crc\" DevicePath \"\"" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.665176 4906 generic.go:334] "Generic (PLEG): container finished" podID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerID="9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371" exitCode=0 Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.665367 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerDied","Data":"9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371"} Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.665542 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4hh7" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.665925 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4hh7" event={"ID":"09b96406-11c2-49bc-a54d-4d68af3cc9ef","Type":"ContainerDied","Data":"9ba744da3451a279c9364ebbb8f22bb5aa42e0b33a222896d4cf67821c068dd4"} Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.665983 4906 scope.go:117] "RemoveContainer" containerID="9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.697437 4906 scope.go:117] "RemoveContainer" containerID="677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.710797 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4hh7"] Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.719722 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v4hh7"] Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.732495 4906 scope.go:117] "RemoveContainer" containerID="908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.760925 4906 scope.go:117] "RemoveContainer" containerID="9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371" Nov 23 08:08:33 crc kubenswrapper[4906]: E1123 08:08:33.762035 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371\": container with ID starting with 9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371 not found: ID does not exist" containerID="9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.762086 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371"} err="failed to get container status \"9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371\": rpc error: code = NotFound desc = could not find container \"9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371\": container with ID starting with 9fc24e3700bf49c13f6b30d3a82c309cbe7e678d87d42a0f31dca2e45bdee371 not found: ID does not exist" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.762120 4906 scope.go:117] "RemoveContainer" containerID="677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012" Nov 23 08:08:33 crc kubenswrapper[4906]: E1123 08:08:33.762667 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012\": container with ID starting with 677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012 not found: ID does not exist" containerID="677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.762936 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012"} err="failed to get container status \"677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012\": rpc error: code = NotFound desc = could not find container \"677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012\": container with ID starting with 677afaa6c33c0b30b77c10d348ab65537589fa483939011820212884118cc012 not found: ID does not exist" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.763099 4906 scope.go:117] "RemoveContainer" containerID="908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8" Nov 23 08:08:33 crc kubenswrapper[4906]: E1123 08:08:33.764046 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8\": container with ID starting with 908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8 not found: ID does not exist" containerID="908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8" Nov 23 08:08:33 crc kubenswrapper[4906]: I1123 08:08:33.764121 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8"} err="failed to get container status \"908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8\": rpc error: code = NotFound desc = could not find container \"908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8\": container with ID starting with 908d3f1e03727b03260c4ff3cb0bf4238f21e37d99e403d7892a61f0fe1c0ce8 not found: ID does not exist" Nov 23 08:08:35 crc kubenswrapper[4906]: I1123 08:08:35.374253 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" path="/var/lib/kubelet/pods/09b96406-11c2-49bc-a54d-4d68af3cc9ef/volumes" Nov 23 08:08:43 crc kubenswrapper[4906]: I1123 08:08:43.357620 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:08:43 crc kubenswrapper[4906]: E1123 08:08:43.358552 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:08:56 crc kubenswrapper[4906]: I1123 08:08:56.357436 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:08:56 crc kubenswrapper[4906]: E1123 08:08:56.358805 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:09:08 crc kubenswrapper[4906]: I1123 08:09:08.356959 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:09:08 crc kubenswrapper[4906]: E1123 08:09:08.357874 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:09:19 crc kubenswrapper[4906]: I1123 08:09:19.357526 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:09:19 crc kubenswrapper[4906]: E1123 08:09:19.359243 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:09:34 crc kubenswrapper[4906]: I1123 08:09:34.357081 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:09:34 crc kubenswrapper[4906]: E1123 08:09:34.358188 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:09:49 crc kubenswrapper[4906]: I1123 08:09:49.357012 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:09:49 crc kubenswrapper[4906]: E1123 08:09:49.357831 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:10:00 crc kubenswrapper[4906]: I1123 08:10:00.356869 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:10:00 crc kubenswrapper[4906]: E1123 08:10:00.357979 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:10:14 crc kubenswrapper[4906]: I1123 08:10:14.357664 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:10:14 crc kubenswrapper[4906]: E1123 08:10:14.358996 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:10:26 crc kubenswrapper[4906]: I1123 08:10:26.357130 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:10:26 crc kubenswrapper[4906]: E1123 08:10:26.358414 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:10:38 crc kubenswrapper[4906]: I1123 08:10:38.357233 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:10:38 crc kubenswrapper[4906]: E1123 08:10:38.358258 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.211313 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7qncl"] Nov 23 08:10:42 crc kubenswrapper[4906]: E1123 08:10:42.212212 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="extract-content" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.212234 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="extract-content" Nov 23 08:10:42 crc kubenswrapper[4906]: E1123 08:10:42.212266 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="registry-server" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.212276 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="registry-server" Nov 23 08:10:42 crc kubenswrapper[4906]: E1123 08:10:42.212302 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="extract-utilities" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.212318 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="extract-utilities" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.212530 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b96406-11c2-49bc-a54d-4d68af3cc9ef" containerName="registry-server" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.213902 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.254251 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qncl"] Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.342477 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-catalog-content\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.342613 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-utilities\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.342653 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4qrv\" (UniqueName: \"kubernetes.io/projected/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-kube-api-access-q4qrv\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.444585 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-catalog-content\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.445488 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-utilities\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.445875 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4qrv\" (UniqueName: \"kubernetes.io/projected/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-kube-api-access-q4qrv\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.445842 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-utilities\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.445390 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-catalog-content\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.470177 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4qrv\" (UniqueName: \"kubernetes.io/projected/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-kube-api-access-q4qrv\") pod \"redhat-marketplace-7qncl\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.545729 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:42 crc kubenswrapper[4906]: I1123 08:10:42.853150 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qncl"] Nov 23 08:10:43 crc kubenswrapper[4906]: I1123 08:10:43.074219 4906 generic.go:334] "Generic (PLEG): container finished" podID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerID="f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a" exitCode=0 Nov 23 08:10:43 crc kubenswrapper[4906]: I1123 08:10:43.074280 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qncl" event={"ID":"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f","Type":"ContainerDied","Data":"f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a"} Nov 23 08:10:43 crc kubenswrapper[4906]: I1123 08:10:43.074354 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qncl" event={"ID":"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f","Type":"ContainerStarted","Data":"5b80fbb3bca34ea5de5d1f0e68ae7505f1e622b83473c62aedd331fc125c3989"} Nov 23 08:10:43 crc kubenswrapper[4906]: I1123 08:10:43.077344 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:10:45 crc kubenswrapper[4906]: I1123 08:10:45.094712 4906 generic.go:334] "Generic (PLEG): container finished" podID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerID="1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed" exitCode=0 Nov 23 08:10:45 crc kubenswrapper[4906]: I1123 08:10:45.095187 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qncl" event={"ID":"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f","Type":"ContainerDied","Data":"1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed"} Nov 23 08:10:46 crc kubenswrapper[4906]: I1123 08:10:46.109730 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qncl" event={"ID":"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f","Type":"ContainerStarted","Data":"66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255"} Nov 23 08:10:46 crc kubenswrapper[4906]: I1123 08:10:46.145468 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7qncl" podStartSLOduration=1.712399688 podStartE2EDuration="4.1454394s" podCreationTimestamp="2025-11-23 08:10:42 +0000 UTC" firstStartedPulling="2025-11-23 08:10:43.077009237 +0000 UTC m=+4858.590400550" lastFinishedPulling="2025-11-23 08:10:45.510048919 +0000 UTC m=+4861.023440262" observedRunningTime="2025-11-23 08:10:46.140971241 +0000 UTC m=+4861.654362554" watchObservedRunningTime="2025-11-23 08:10:46.1454394 +0000 UTC m=+4861.658830713" Nov 23 08:10:52 crc kubenswrapper[4906]: I1123 08:10:52.356584 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:10:52 crc kubenswrapper[4906]: I1123 08:10:52.546738 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:52 crc kubenswrapper[4906]: I1123 08:10:52.547196 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:52 crc kubenswrapper[4906]: I1123 08:10:52.601874 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:53 crc kubenswrapper[4906]: I1123 08:10:53.183882 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"3b67e657e617cce35f5b2a269e0dce71c16b1000fae65b2dee5c28285fe1d906"} Nov 23 08:10:53 crc kubenswrapper[4906]: I1123 08:10:53.261072 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:53 crc kubenswrapper[4906]: I1123 08:10:53.313256 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qncl"] Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.206446 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7qncl" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="registry-server" containerID="cri-o://66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255" gracePeriod=2 Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.883607 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.924474 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4qrv\" (UniqueName: \"kubernetes.io/projected/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-kube-api-access-q4qrv\") pod \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.924625 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-utilities\") pod \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.924769 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-catalog-content\") pod \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\" (UID: \"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f\") " Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.926118 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-utilities" (OuterVolumeSpecName: "utilities") pod "ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" (UID: "ad28e1a6-8411-4a8f-8b55-a30fea3bac2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.934441 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-kube-api-access-q4qrv" (OuterVolumeSpecName: "kube-api-access-q4qrv") pod "ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" (UID: "ad28e1a6-8411-4a8f-8b55-a30fea3bac2f"). InnerVolumeSpecName "kube-api-access-q4qrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:10:55 crc kubenswrapper[4906]: I1123 08:10:55.963524 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" (UID: "ad28e1a6-8411-4a8f-8b55-a30fea3bac2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.027294 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4qrv\" (UniqueName: \"kubernetes.io/projected/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-kube-api-access-q4qrv\") on node \"crc\" DevicePath \"\"" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.027343 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.027354 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.221031 4906 generic.go:334] "Generic (PLEG): container finished" podID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerID="66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255" exitCode=0 Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.221118 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qncl" event={"ID":"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f","Type":"ContainerDied","Data":"66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255"} Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.221141 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qncl" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.221183 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qncl" event={"ID":"ad28e1a6-8411-4a8f-8b55-a30fea3bac2f","Type":"ContainerDied","Data":"5b80fbb3bca34ea5de5d1f0e68ae7505f1e622b83473c62aedd331fc125c3989"} Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.221226 4906 scope.go:117] "RemoveContainer" containerID="66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.267439 4906 scope.go:117] "RemoveContainer" containerID="1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.269085 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qncl"] Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.298981 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qncl"] Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.326526 4906 scope.go:117] "RemoveContainer" containerID="f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.355909 4906 scope.go:117] "RemoveContainer" containerID="66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255" Nov 23 08:10:56 crc kubenswrapper[4906]: E1123 08:10:56.357083 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255\": container with ID starting with 66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255 not found: ID does not exist" containerID="66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.357155 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255"} err="failed to get container status \"66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255\": rpc error: code = NotFound desc = could not find container \"66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255\": container with ID starting with 66537f437515bfaa3b6ad15708bb19aa625a1398176546bf5e0a5948fef05255 not found: ID does not exist" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.357196 4906 scope.go:117] "RemoveContainer" containerID="1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed" Nov 23 08:10:56 crc kubenswrapper[4906]: E1123 08:10:56.358208 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed\": container with ID starting with 1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed not found: ID does not exist" containerID="1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.358255 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed"} err="failed to get container status \"1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed\": rpc error: code = NotFound desc = could not find container \"1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed\": container with ID starting with 1dbc79c6d13fb912d4a454d0034f349f4b65e0d3ee401be7d67f2c13bbb100ed not found: ID does not exist" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.358285 4906 scope.go:117] "RemoveContainer" containerID="f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a" Nov 23 08:10:56 crc kubenswrapper[4906]: E1123 08:10:56.359538 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a\": container with ID starting with f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a not found: ID does not exist" containerID="f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a" Nov 23 08:10:56 crc kubenswrapper[4906]: I1123 08:10:56.359556 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a"} err="failed to get container status \"f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a\": rpc error: code = NotFound desc = could not find container \"f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a\": container with ID starting with f54b302fbcca1765f6fbb149233810084baa3960567d2bc528731a069bd9e56a not found: ID does not exist" Nov 23 08:10:57 crc kubenswrapper[4906]: I1123 08:10:57.369704 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" path="/var/lib/kubelet/pods/ad28e1a6-8411-4a8f-8b55-a30fea3bac2f/volumes" Nov 23 08:13:20 crc kubenswrapper[4906]: I1123 08:13:20.945749 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:13:20 crc kubenswrapper[4906]: I1123 08:13:20.946508 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:13:50 crc kubenswrapper[4906]: I1123 08:13:50.945733 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:13:50 crc kubenswrapper[4906]: I1123 08:13:50.947837 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:14:20 crc kubenswrapper[4906]: I1123 08:14:20.945773 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:14:20 crc kubenswrapper[4906]: I1123 08:14:20.946397 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:14:20 crc kubenswrapper[4906]: I1123 08:14:20.946497 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:14:20 crc kubenswrapper[4906]: I1123 08:14:20.947393 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3b67e657e617cce35f5b2a269e0dce71c16b1000fae65b2dee5c28285fe1d906"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:14:20 crc kubenswrapper[4906]: I1123 08:14:20.947493 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://3b67e657e617cce35f5b2a269e0dce71c16b1000fae65b2dee5c28285fe1d906" gracePeriod=600 Nov 23 08:14:21 crc kubenswrapper[4906]: I1123 08:14:21.248236 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="3b67e657e617cce35f5b2a269e0dce71c16b1000fae65b2dee5c28285fe1d906" exitCode=0 Nov 23 08:14:21 crc kubenswrapper[4906]: I1123 08:14:21.248309 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"3b67e657e617cce35f5b2a269e0dce71c16b1000fae65b2dee5c28285fe1d906"} Nov 23 08:14:21 crc kubenswrapper[4906]: I1123 08:14:21.248627 4906 scope.go:117] "RemoveContainer" containerID="aafbbca6420f92478538c6f6db24e4afe2c358bf4bc7641c844bf9a932975f45" Nov 23 08:14:22 crc kubenswrapper[4906]: I1123 08:14:22.261893 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6"} Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.202758 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45"] Nov 23 08:15:00 crc kubenswrapper[4906]: E1123 08:15:00.205027 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="registry-server" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.205069 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="registry-server" Nov 23 08:15:00 crc kubenswrapper[4906]: E1123 08:15:00.205152 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="extract-utilities" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.205161 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="extract-utilities" Nov 23 08:15:00 crc kubenswrapper[4906]: E1123 08:15:00.205208 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="extract-content" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.205215 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="extract-content" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.208802 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad28e1a6-8411-4a8f-8b55-a30fea3bac2f" containerName="registry-server" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.210268 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.221094 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.221932 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.222789 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45"] Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.272888 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-secret-volume\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.272967 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-config-volume\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.273005 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvfmn\" (UniqueName: \"kubernetes.io/projected/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-kube-api-access-rvfmn\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.373937 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-secret-volume\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.374007 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-config-volume\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.374031 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvfmn\" (UniqueName: \"kubernetes.io/projected/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-kube-api-access-rvfmn\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.375000 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-config-volume\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.388335 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-secret-volume\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.388714 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvfmn\" (UniqueName: \"kubernetes.io/projected/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-kube-api-access-rvfmn\") pod \"collect-profiles-29398095-86c45\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.555482 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:00 crc kubenswrapper[4906]: I1123 08:15:00.994822 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45"] Nov 23 08:15:01 crc kubenswrapper[4906]: I1123 08:15:01.618624 4906 generic.go:334] "Generic (PLEG): container finished" podID="f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" containerID="fb87792ee365d7e810b540eee0d4886bfa04b51e9f9694dcfa762128371b778d" exitCode=0 Nov 23 08:15:01 crc kubenswrapper[4906]: I1123 08:15:01.618722 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" event={"ID":"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8","Type":"ContainerDied","Data":"fb87792ee365d7e810b540eee0d4886bfa04b51e9f9694dcfa762128371b778d"} Nov 23 08:15:01 crc kubenswrapper[4906]: I1123 08:15:01.619103 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" event={"ID":"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8","Type":"ContainerStarted","Data":"0d35b5a8a31ffe176bf9c8691a4aabf69329ccad3387c0cb06a11f88d00680a6"} Nov 23 08:15:02 crc kubenswrapper[4906]: I1123 08:15:02.978066 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.116499 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-secret-volume\") pod \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.116794 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvfmn\" (UniqueName: \"kubernetes.io/projected/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-kube-api-access-rvfmn\") pod \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.116901 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-config-volume\") pod \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\" (UID: \"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8\") " Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.117778 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-config-volume" (OuterVolumeSpecName: "config-volume") pod "f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" (UID: "f6efcfd7-e3ef-404f-9a43-8cb8514e43b8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.125035 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" (UID: "f6efcfd7-e3ef-404f-9a43-8cb8514e43b8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.125078 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-kube-api-access-rvfmn" (OuterVolumeSpecName: "kube-api-access-rvfmn") pod "f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" (UID: "f6efcfd7-e3ef-404f-9a43-8cb8514e43b8"). InnerVolumeSpecName "kube-api-access-rvfmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.219754 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.219810 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.219828 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvfmn\" (UniqueName: \"kubernetes.io/projected/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8-kube-api-access-rvfmn\") on node \"crc\" DevicePath \"\"" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.638501 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" event={"ID":"f6efcfd7-e3ef-404f-9a43-8cb8514e43b8","Type":"ContainerDied","Data":"0d35b5a8a31ffe176bf9c8691a4aabf69329ccad3387c0cb06a11f88d00680a6"} Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.638585 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d35b5a8a31ffe176bf9c8691a4aabf69329ccad3387c0cb06a11f88d00680a6" Nov 23 08:15:03 crc kubenswrapper[4906]: I1123 08:15:03.638600 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45" Nov 23 08:15:04 crc kubenswrapper[4906]: I1123 08:15:04.074332 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms"] Nov 23 08:15:04 crc kubenswrapper[4906]: I1123 08:15:04.080996 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-ds2ms"] Nov 23 08:15:05 crc kubenswrapper[4906]: I1123 08:15:05.375111 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80abab3e-830e-493f-8bb1-6bb9d8bd1a32" path="/var/lib/kubelet/pods/80abab3e-830e-493f-8bb1-6bb9d8bd1a32/volumes" Nov 23 08:15:58 crc kubenswrapper[4906]: I1123 08:15:58.267443 4906 scope.go:117] "RemoveContainer" containerID="df20c20b9d37b1a5a5acf67ef09647ddf26fa72f65a94ae13f902a158708d4c1" Nov 23 08:16:50 crc kubenswrapper[4906]: I1123 08:16:50.946229 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:16:50 crc kubenswrapper[4906]: I1123 08:16:50.947245 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.833480 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-72htx"] Nov 23 08:16:51 crc kubenswrapper[4906]: E1123 08:16:51.836196 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" containerName="collect-profiles" Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.836260 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" containerName="collect-profiles" Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.838535 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" containerName="collect-profiles" Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.846474 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.875468 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-72htx"] Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.956590 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-catalog-content\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.957802 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-utilities\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:51 crc kubenswrapper[4906]: I1123 08:16:51.958013 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9xf4\" (UniqueName: \"kubernetes.io/projected/0823fed6-81b3-4cb6-a210-0a818d8e47d9-kube-api-access-r9xf4\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.059557 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9xf4\" (UniqueName: \"kubernetes.io/projected/0823fed6-81b3-4cb6-a210-0a818d8e47d9-kube-api-access-r9xf4\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.059751 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-catalog-content\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.059831 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-utilities\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.060622 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-catalog-content\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.060743 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-utilities\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.097573 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9xf4\" (UniqueName: \"kubernetes.io/projected/0823fed6-81b3-4cb6-a210-0a818d8e47d9-kube-api-access-r9xf4\") pod \"redhat-operators-72htx\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.187381 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.455583 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-72htx"] Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.743614 4906 generic.go:334] "Generic (PLEG): container finished" podID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerID="0fad7a4879d40dc4bf76efa52904e797593fc3e692a20f8b1f337d12153ebc85" exitCode=0 Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.743781 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72htx" event={"ID":"0823fed6-81b3-4cb6-a210-0a818d8e47d9","Type":"ContainerDied","Data":"0fad7a4879d40dc4bf76efa52904e797593fc3e692a20f8b1f337d12153ebc85"} Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.744009 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72htx" event={"ID":"0823fed6-81b3-4cb6-a210-0a818d8e47d9","Type":"ContainerStarted","Data":"0c20dcc1c2e28189e50ab01d5aa5796db788aa9c84c04d0c251331fa8ce6913c"} Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.745562 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.817090 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jjqwv"] Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.818587 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.828851 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jjqwv"] Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.975049 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-catalog-content\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.975115 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-utilities\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:52 crc kubenswrapper[4906]: I1123 08:16:52.975189 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxvsc\" (UniqueName: \"kubernetes.io/projected/275fc6bf-f979-41ff-8896-675d176df766-kube-api-access-lxvsc\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.076508 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-catalog-content\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.076575 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-utilities\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.076617 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxvsc\" (UniqueName: \"kubernetes.io/projected/275fc6bf-f979-41ff-8896-675d176df766-kube-api-access-lxvsc\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.077016 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-catalog-content\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.077110 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-utilities\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.096160 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxvsc\" (UniqueName: \"kubernetes.io/projected/275fc6bf-f979-41ff-8896-675d176df766-kube-api-access-lxvsc\") pod \"community-operators-jjqwv\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.182656 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.537984 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jjqwv"] Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.752318 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72htx" event={"ID":"0823fed6-81b3-4cb6-a210-0a818d8e47d9","Type":"ContainerStarted","Data":"28b96ecb7122f4d93a2671defcc56a1a06157be2720bbd828171c63b21193f3a"} Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.753775 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerStarted","Data":"466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2"} Nov 23 08:16:53 crc kubenswrapper[4906]: I1123 08:16:53.753818 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerStarted","Data":"a5d7597cc77ecfc1c05e163cf8efc11aeee840a37c9d80795c765c6d72a044de"} Nov 23 08:16:54 crc kubenswrapper[4906]: I1123 08:16:54.765086 4906 generic.go:334] "Generic (PLEG): container finished" podID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerID="28b96ecb7122f4d93a2671defcc56a1a06157be2720bbd828171c63b21193f3a" exitCode=0 Nov 23 08:16:54 crc kubenswrapper[4906]: I1123 08:16:54.765157 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72htx" event={"ID":"0823fed6-81b3-4cb6-a210-0a818d8e47d9","Type":"ContainerDied","Data":"28b96ecb7122f4d93a2671defcc56a1a06157be2720bbd828171c63b21193f3a"} Nov 23 08:16:54 crc kubenswrapper[4906]: I1123 08:16:54.766729 4906 generic.go:334] "Generic (PLEG): container finished" podID="275fc6bf-f979-41ff-8896-675d176df766" containerID="466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2" exitCode=0 Nov 23 08:16:54 crc kubenswrapper[4906]: I1123 08:16:54.766790 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerDied","Data":"466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2"} Nov 23 08:16:55 crc kubenswrapper[4906]: I1123 08:16:55.784792 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72htx" event={"ID":"0823fed6-81b3-4cb6-a210-0a818d8e47d9","Type":"ContainerStarted","Data":"054d045b4d855bfae3baa9f897f76da55fcea6918a825ef09d85eaa68fe4f735"} Nov 23 08:16:55 crc kubenswrapper[4906]: I1123 08:16:55.789241 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerStarted","Data":"7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0"} Nov 23 08:16:55 crc kubenswrapper[4906]: I1123 08:16:55.807421 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-72htx" podStartSLOduration=2.361285582 podStartE2EDuration="4.807389529s" podCreationTimestamp="2025-11-23 08:16:51 +0000 UTC" firstStartedPulling="2025-11-23 08:16:52.745313017 +0000 UTC m=+5228.258704320" lastFinishedPulling="2025-11-23 08:16:55.191416954 +0000 UTC m=+5230.704808267" observedRunningTime="2025-11-23 08:16:55.801429202 +0000 UTC m=+5231.314820515" watchObservedRunningTime="2025-11-23 08:16:55.807389529 +0000 UTC m=+5231.320780842" Nov 23 08:16:56 crc kubenswrapper[4906]: I1123 08:16:56.799779 4906 generic.go:334] "Generic (PLEG): container finished" podID="275fc6bf-f979-41ff-8896-675d176df766" containerID="7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0" exitCode=0 Nov 23 08:16:56 crc kubenswrapper[4906]: I1123 08:16:56.799897 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerDied","Data":"7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0"} Nov 23 08:16:57 crc kubenswrapper[4906]: I1123 08:16:57.806872 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerStarted","Data":"477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75"} Nov 23 08:16:57 crc kubenswrapper[4906]: I1123 08:16:57.826781 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jjqwv" podStartSLOduration=3.394349201 podStartE2EDuration="5.826763298s" podCreationTimestamp="2025-11-23 08:16:52 +0000 UTC" firstStartedPulling="2025-11-23 08:16:54.767967203 +0000 UTC m=+5230.281358506" lastFinishedPulling="2025-11-23 08:16:57.20038126 +0000 UTC m=+5232.713772603" observedRunningTime="2025-11-23 08:16:57.824361105 +0000 UTC m=+5233.337752408" watchObservedRunningTime="2025-11-23 08:16:57.826763298 +0000 UTC m=+5233.340154601" Nov 23 08:17:02 crc kubenswrapper[4906]: I1123 08:17:02.187811 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:17:02 crc kubenswrapper[4906]: I1123 08:17:02.189662 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:17:02 crc kubenswrapper[4906]: I1123 08:17:02.251795 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:17:02 crc kubenswrapper[4906]: I1123 08:17:02.923525 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:17:02 crc kubenswrapper[4906]: I1123 08:17:02.984658 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-72htx"] Nov 23 08:17:03 crc kubenswrapper[4906]: I1123 08:17:03.183818 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:17:03 crc kubenswrapper[4906]: I1123 08:17:03.184387 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:17:03 crc kubenswrapper[4906]: I1123 08:17:03.246757 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:17:04 crc kubenswrapper[4906]: I1123 08:17:03.929118 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:17:04 crc kubenswrapper[4906]: I1123 08:17:04.869841 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-72htx" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="registry-server" containerID="cri-o://054d045b4d855bfae3baa9f897f76da55fcea6918a825ef09d85eaa68fe4f735" gracePeriod=2 Nov 23 08:17:05 crc kubenswrapper[4906]: I1123 08:17:05.017156 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jjqwv"] Nov 23 08:17:05 crc kubenswrapper[4906]: I1123 08:17:05.875843 4906 generic.go:334] "Generic (PLEG): container finished" podID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerID="054d045b4d855bfae3baa9f897f76da55fcea6918a825ef09d85eaa68fe4f735" exitCode=0 Nov 23 08:17:05 crc kubenswrapper[4906]: I1123 08:17:05.876549 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72htx" event={"ID":"0823fed6-81b3-4cb6-a210-0a818d8e47d9","Type":"ContainerDied","Data":"054d045b4d855bfae3baa9f897f76da55fcea6918a825ef09d85eaa68fe4f735"} Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.419861 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.511614 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-catalog-content\") pod \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.511741 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-utilities\") pod \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.511798 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9xf4\" (UniqueName: \"kubernetes.io/projected/0823fed6-81b3-4cb6-a210-0a818d8e47d9-kube-api-access-r9xf4\") pod \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\" (UID: \"0823fed6-81b3-4cb6-a210-0a818d8e47d9\") " Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.512968 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-utilities" (OuterVolumeSpecName: "utilities") pod "0823fed6-81b3-4cb6-a210-0a818d8e47d9" (UID: "0823fed6-81b3-4cb6-a210-0a818d8e47d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.520556 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0823fed6-81b3-4cb6-a210-0a818d8e47d9-kube-api-access-r9xf4" (OuterVolumeSpecName: "kube-api-access-r9xf4") pod "0823fed6-81b3-4cb6-a210-0a818d8e47d9" (UID: "0823fed6-81b3-4cb6-a210-0a818d8e47d9"). InnerVolumeSpecName "kube-api-access-r9xf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.613357 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.613394 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9xf4\" (UniqueName: \"kubernetes.io/projected/0823fed6-81b3-4cb6-a210-0a818d8e47d9-kube-api-access-r9xf4\") on node \"crc\" DevicePath \"\"" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.621828 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0823fed6-81b3-4cb6-a210-0a818d8e47d9" (UID: "0823fed6-81b3-4cb6-a210-0a818d8e47d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.716031 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0823fed6-81b3-4cb6-a210-0a818d8e47d9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.886357 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72htx" event={"ID":"0823fed6-81b3-4cb6-a210-0a818d8e47d9","Type":"ContainerDied","Data":"0c20dcc1c2e28189e50ab01d5aa5796db788aa9c84c04d0c251331fa8ce6913c"} Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.886404 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72htx" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.886456 4906 scope.go:117] "RemoveContainer" containerID="054d045b4d855bfae3baa9f897f76da55fcea6918a825ef09d85eaa68fe4f735" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.886690 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jjqwv" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="registry-server" containerID="cri-o://477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75" gracePeriod=2 Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.905944 4906 scope.go:117] "RemoveContainer" containerID="28b96ecb7122f4d93a2671defcc56a1a06157be2720bbd828171c63b21193f3a" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.944607 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-72htx"] Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.945529 4906 scope.go:117] "RemoveContainer" containerID="0fad7a4879d40dc4bf76efa52904e797593fc3e692a20f8b1f337d12153ebc85" Nov 23 08:17:06 crc kubenswrapper[4906]: I1123 08:17:06.950443 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-72htx"] Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.285338 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.371919 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" path="/var/lib/kubelet/pods/0823fed6-81b3-4cb6-a210-0a818d8e47d9/volumes" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.426578 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-utilities\") pod \"275fc6bf-f979-41ff-8896-675d176df766\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.426781 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxvsc\" (UniqueName: \"kubernetes.io/projected/275fc6bf-f979-41ff-8896-675d176df766-kube-api-access-lxvsc\") pod \"275fc6bf-f979-41ff-8896-675d176df766\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.426893 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-catalog-content\") pod \"275fc6bf-f979-41ff-8896-675d176df766\" (UID: \"275fc6bf-f979-41ff-8896-675d176df766\") " Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.429012 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-utilities" (OuterVolumeSpecName: "utilities") pod "275fc6bf-f979-41ff-8896-675d176df766" (UID: "275fc6bf-f979-41ff-8896-675d176df766"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.433715 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/275fc6bf-f979-41ff-8896-675d176df766-kube-api-access-lxvsc" (OuterVolumeSpecName: "kube-api-access-lxvsc") pod "275fc6bf-f979-41ff-8896-675d176df766" (UID: "275fc6bf-f979-41ff-8896-675d176df766"). InnerVolumeSpecName "kube-api-access-lxvsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.502757 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "275fc6bf-f979-41ff-8896-675d176df766" (UID: "275fc6bf-f979-41ff-8896-675d176df766"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.531562 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.532031 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxvsc\" (UniqueName: \"kubernetes.io/projected/275fc6bf-f979-41ff-8896-675d176df766-kube-api-access-lxvsc\") on node \"crc\" DevicePath \"\"" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.532167 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275fc6bf-f979-41ff-8896-675d176df766-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.900497 4906 generic.go:334] "Generic (PLEG): container finished" podID="275fc6bf-f979-41ff-8896-675d176df766" containerID="477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75" exitCode=0 Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.900582 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerDied","Data":"477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75"} Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.901719 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjqwv" event={"ID":"275fc6bf-f979-41ff-8896-675d176df766","Type":"ContainerDied","Data":"a5d7597cc77ecfc1c05e163cf8efc11aeee840a37c9d80795c765c6d72a044de"} Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.900619 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjqwv" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.901769 4906 scope.go:117] "RemoveContainer" containerID="477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.936517 4906 scope.go:117] "RemoveContainer" containerID="7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.939211 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jjqwv"] Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.946366 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jjqwv"] Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.960281 4906 scope.go:117] "RemoveContainer" containerID="466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.988259 4906 scope.go:117] "RemoveContainer" containerID="477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75" Nov 23 08:17:07 crc kubenswrapper[4906]: E1123 08:17:07.988779 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75\": container with ID starting with 477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75 not found: ID does not exist" containerID="477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.988819 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75"} err="failed to get container status \"477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75\": rpc error: code = NotFound desc = could not find container \"477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75\": container with ID starting with 477888bdd9edc35f3933e0f467e23ea95422ad2a715f471e0818c66fce738b75 not found: ID does not exist" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.988851 4906 scope.go:117] "RemoveContainer" containerID="7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0" Nov 23 08:17:07 crc kubenswrapper[4906]: E1123 08:17:07.989158 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0\": container with ID starting with 7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0 not found: ID does not exist" containerID="7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.989198 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0"} err="failed to get container status \"7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0\": rpc error: code = NotFound desc = could not find container \"7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0\": container with ID starting with 7287aa36aaa4f7a998e48114449bbfb09eceeba6fe3266ce116faa220a88dac0 not found: ID does not exist" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.989218 4906 scope.go:117] "RemoveContainer" containerID="466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2" Nov 23 08:17:07 crc kubenswrapper[4906]: E1123 08:17:07.989573 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2\": container with ID starting with 466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2 not found: ID does not exist" containerID="466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2" Nov 23 08:17:07 crc kubenswrapper[4906]: I1123 08:17:07.989728 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2"} err="failed to get container status \"466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2\": rpc error: code = NotFound desc = could not find container \"466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2\": container with ID starting with 466cc43f41a44dfc713553347bf8efdbbdd391557f247283cc1022eb342bf5f2 not found: ID does not exist" Nov 23 08:17:09 crc kubenswrapper[4906]: I1123 08:17:09.370167 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="275fc6bf-f979-41ff-8896-675d176df766" path="/var/lib/kubelet/pods/275fc6bf-f979-41ff-8896-675d176df766/volumes" Nov 23 08:17:20 crc kubenswrapper[4906]: I1123 08:17:20.945622 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:17:20 crc kubenswrapper[4906]: I1123 08:17:20.947259 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:17:50 crc kubenswrapper[4906]: I1123 08:17:50.946263 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:17:50 crc kubenswrapper[4906]: I1123 08:17:50.947018 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:17:50 crc kubenswrapper[4906]: I1123 08:17:50.947070 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:17:50 crc kubenswrapper[4906]: I1123 08:17:50.947764 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:17:50 crc kubenswrapper[4906]: I1123 08:17:50.947835 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" gracePeriod=600 Nov 23 08:17:51 crc kubenswrapper[4906]: E1123 08:17:51.084525 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:17:51 crc kubenswrapper[4906]: I1123 08:17:51.345659 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" exitCode=0 Nov 23 08:17:51 crc kubenswrapper[4906]: I1123 08:17:51.345723 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6"} Nov 23 08:17:51 crc kubenswrapper[4906]: I1123 08:17:51.346209 4906 scope.go:117] "RemoveContainer" containerID="3b67e657e617cce35f5b2a269e0dce71c16b1000fae65b2dee5c28285fe1d906" Nov 23 08:17:51 crc kubenswrapper[4906]: I1123 08:17:51.347376 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:17:51 crc kubenswrapper[4906]: E1123 08:17:51.347943 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:18:06 crc kubenswrapper[4906]: I1123 08:18:06.357976 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:18:06 crc kubenswrapper[4906]: E1123 08:18:06.359581 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:18:19 crc kubenswrapper[4906]: I1123 08:18:19.356629 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:18:19 crc kubenswrapper[4906]: E1123 08:18:19.357448 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:18:30 crc kubenswrapper[4906]: I1123 08:18:30.356789 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:18:30 crc kubenswrapper[4906]: E1123 08:18:30.357746 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:18:44 crc kubenswrapper[4906]: I1123 08:18:44.358641 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:18:44 crc kubenswrapper[4906]: E1123 08:18:44.360003 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.484444 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9wpk5"] Nov 23 08:18:54 crc kubenswrapper[4906]: E1123 08:18:54.485802 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="extract-utilities" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.485835 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="extract-utilities" Nov 23 08:18:54 crc kubenswrapper[4906]: E1123 08:18:54.485990 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="extract-utilities" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.486010 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="extract-utilities" Nov 23 08:18:54 crc kubenswrapper[4906]: E1123 08:18:54.486122 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="extract-content" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.486145 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="extract-content" Nov 23 08:18:54 crc kubenswrapper[4906]: E1123 08:18:54.486166 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="extract-content" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.486181 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="extract-content" Nov 23 08:18:54 crc kubenswrapper[4906]: E1123 08:18:54.486206 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="registry-server" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.486222 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="registry-server" Nov 23 08:18:54 crc kubenswrapper[4906]: E1123 08:18:54.486256 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="registry-server" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.486272 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="registry-server" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.486663 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="0823fed6-81b3-4cb6-a210-0a818d8e47d9" containerName="registry-server" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.486862 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="275fc6bf-f979-41ff-8896-675d176df766" containerName="registry-server" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.489230 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.495592 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9wpk5"] Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.649076 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-utilities\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.649136 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-catalog-content\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.649243 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9lm6\" (UniqueName: \"kubernetes.io/projected/fdd5d275-a1c6-45af-832b-d511a444b3b8-kube-api-access-v9lm6\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.750091 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9lm6\" (UniqueName: \"kubernetes.io/projected/fdd5d275-a1c6-45af-832b-d511a444b3b8-kube-api-access-v9lm6\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.750176 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-utilities\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.750205 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-catalog-content\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.750693 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-catalog-content\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.750744 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-utilities\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.780816 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9lm6\" (UniqueName: \"kubernetes.io/projected/fdd5d275-a1c6-45af-832b-d511a444b3b8-kube-api-access-v9lm6\") pod \"certified-operators-9wpk5\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:54 crc kubenswrapper[4906]: I1123 08:18:54.822846 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:18:55 crc kubenswrapper[4906]: I1123 08:18:55.329946 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9wpk5"] Nov 23 08:18:55 crc kubenswrapper[4906]: I1123 08:18:55.992587 4906 generic.go:334] "Generic (PLEG): container finished" podID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerID="c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e" exitCode=0 Nov 23 08:18:55 crc kubenswrapper[4906]: I1123 08:18:55.992732 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wpk5" event={"ID":"fdd5d275-a1c6-45af-832b-d511a444b3b8","Type":"ContainerDied","Data":"c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e"} Nov 23 08:18:55 crc kubenswrapper[4906]: I1123 08:18:55.992841 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wpk5" event={"ID":"fdd5d275-a1c6-45af-832b-d511a444b3b8","Type":"ContainerStarted","Data":"5dac7e8d753501569ae3ccadb047d9e6ceff2d330d7563a2350d14ef218d11e0"} Nov 23 08:18:57 crc kubenswrapper[4906]: I1123 08:18:57.009050 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wpk5" event={"ID":"fdd5d275-a1c6-45af-832b-d511a444b3b8","Type":"ContainerStarted","Data":"b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134"} Nov 23 08:18:57 crc kubenswrapper[4906]: I1123 08:18:57.357544 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:18:57 crc kubenswrapper[4906]: E1123 08:18:57.357832 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:18:58 crc kubenswrapper[4906]: I1123 08:18:58.019221 4906 generic.go:334] "Generic (PLEG): container finished" podID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerID="b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134" exitCode=0 Nov 23 08:18:58 crc kubenswrapper[4906]: I1123 08:18:58.019295 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wpk5" event={"ID":"fdd5d275-a1c6-45af-832b-d511a444b3b8","Type":"ContainerDied","Data":"b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134"} Nov 23 08:18:59 crc kubenswrapper[4906]: I1123 08:18:59.031378 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wpk5" event={"ID":"fdd5d275-a1c6-45af-832b-d511a444b3b8","Type":"ContainerStarted","Data":"b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6"} Nov 23 08:18:59 crc kubenswrapper[4906]: I1123 08:18:59.069047 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9wpk5" podStartSLOduration=2.6648990489999997 podStartE2EDuration="5.069030179s" podCreationTimestamp="2025-11-23 08:18:54 +0000 UTC" firstStartedPulling="2025-11-23 08:18:55.995239527 +0000 UTC m=+5351.508630870" lastFinishedPulling="2025-11-23 08:18:58.399370697 +0000 UTC m=+5353.912762000" observedRunningTime="2025-11-23 08:18:59.059717013 +0000 UTC m=+5354.573108336" watchObservedRunningTime="2025-11-23 08:18:59.069030179 +0000 UTC m=+5354.582421482" Nov 23 08:19:04 crc kubenswrapper[4906]: I1123 08:19:04.823941 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:19:04 crc kubenswrapper[4906]: I1123 08:19:04.824906 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:19:04 crc kubenswrapper[4906]: I1123 08:19:04.896117 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:19:05 crc kubenswrapper[4906]: I1123 08:19:05.147666 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:19:05 crc kubenswrapper[4906]: I1123 08:19:05.219596 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9wpk5"] Nov 23 08:19:07 crc kubenswrapper[4906]: I1123 08:19:07.092815 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9wpk5" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="registry-server" containerID="cri-o://b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6" gracePeriod=2 Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.043909 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.101491 4906 generic.go:334] "Generic (PLEG): container finished" podID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerID="b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6" exitCode=0 Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.101543 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wpk5" event={"ID":"fdd5d275-a1c6-45af-832b-d511a444b3b8","Type":"ContainerDied","Data":"b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6"} Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.101606 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wpk5" event={"ID":"fdd5d275-a1c6-45af-832b-d511a444b3b8","Type":"ContainerDied","Data":"5dac7e8d753501569ae3ccadb047d9e6ceff2d330d7563a2350d14ef218d11e0"} Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.101625 4906 scope.go:117] "RemoveContainer" containerID="b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.102763 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wpk5" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.120311 4906 scope.go:117] "RemoveContainer" containerID="b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.137751 4906 scope.go:117] "RemoveContainer" containerID="c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.159050 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9lm6\" (UniqueName: \"kubernetes.io/projected/fdd5d275-a1c6-45af-832b-d511a444b3b8-kube-api-access-v9lm6\") pod \"fdd5d275-a1c6-45af-832b-d511a444b3b8\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.159168 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-catalog-content\") pod \"fdd5d275-a1c6-45af-832b-d511a444b3b8\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.159274 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-utilities\") pod \"fdd5d275-a1c6-45af-832b-d511a444b3b8\" (UID: \"fdd5d275-a1c6-45af-832b-d511a444b3b8\") " Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.160312 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-utilities" (OuterVolumeSpecName: "utilities") pod "fdd5d275-a1c6-45af-832b-d511a444b3b8" (UID: "fdd5d275-a1c6-45af-832b-d511a444b3b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.163006 4906 scope.go:117] "RemoveContainer" containerID="b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.166815 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdd5d275-a1c6-45af-832b-d511a444b3b8-kube-api-access-v9lm6" (OuterVolumeSpecName: "kube-api-access-v9lm6") pod "fdd5d275-a1c6-45af-832b-d511a444b3b8" (UID: "fdd5d275-a1c6-45af-832b-d511a444b3b8"). InnerVolumeSpecName "kube-api-access-v9lm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:19:08 crc kubenswrapper[4906]: E1123 08:19:08.167139 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6\": container with ID starting with b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6 not found: ID does not exist" containerID="b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.167231 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6"} err="failed to get container status \"b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6\": rpc error: code = NotFound desc = could not find container \"b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6\": container with ID starting with b6ccb0f48e170bad6b5ed27dc2b4bbc02c404eebcc341d818b085177322491a6 not found: ID does not exist" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.167327 4906 scope.go:117] "RemoveContainer" containerID="b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134" Nov 23 08:19:08 crc kubenswrapper[4906]: E1123 08:19:08.167781 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134\": container with ID starting with b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134 not found: ID does not exist" containerID="b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.167872 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134"} err="failed to get container status \"b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134\": rpc error: code = NotFound desc = could not find container \"b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134\": container with ID starting with b4dab70656b9dd81ce19cdb3fb823a6b68510fa07692f91ccc06e969fe252134 not found: ID does not exist" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.167946 4906 scope.go:117] "RemoveContainer" containerID="c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e" Nov 23 08:19:08 crc kubenswrapper[4906]: E1123 08:19:08.168221 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e\": container with ID starting with c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e not found: ID does not exist" containerID="c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.168308 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e"} err="failed to get container status \"c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e\": rpc error: code = NotFound desc = could not find container \"c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e\": container with ID starting with c3a65e8851bf2ab00fc953374ef863568a2e18fcc706b9d3ba9472d5f4fce48e not found: ID does not exist" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.208588 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdd5d275-a1c6-45af-832b-d511a444b3b8" (UID: "fdd5d275-a1c6-45af-832b-d511a444b3b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.261016 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.261326 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd5d275-a1c6-45af-832b-d511a444b3b8-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.261400 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9lm6\" (UniqueName: \"kubernetes.io/projected/fdd5d275-a1c6-45af-832b-d511a444b3b8-kube-api-access-v9lm6\") on node \"crc\" DevicePath \"\"" Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.434093 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9wpk5"] Nov 23 08:19:08 crc kubenswrapper[4906]: I1123 08:19:08.440647 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9wpk5"] Nov 23 08:19:09 crc kubenswrapper[4906]: I1123 08:19:09.356651 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:19:09 crc kubenswrapper[4906]: E1123 08:19:09.357213 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:19:09 crc kubenswrapper[4906]: I1123 08:19:09.368482 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" path="/var/lib/kubelet/pods/fdd5d275-a1c6-45af-832b-d511a444b3b8/volumes" Nov 23 08:19:23 crc kubenswrapper[4906]: I1123 08:19:23.357557 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:19:23 crc kubenswrapper[4906]: E1123 08:19:23.358851 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:19:37 crc kubenswrapper[4906]: I1123 08:19:37.357290 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:19:37 crc kubenswrapper[4906]: E1123 08:19:37.358151 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:19:49 crc kubenswrapper[4906]: I1123 08:19:49.356380 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:19:49 crc kubenswrapper[4906]: E1123 08:19:49.357267 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:20:02 crc kubenswrapper[4906]: I1123 08:20:02.357054 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:20:02 crc kubenswrapper[4906]: E1123 08:20:02.357879 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:20:14 crc kubenswrapper[4906]: I1123 08:20:14.356992 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:20:14 crc kubenswrapper[4906]: E1123 08:20:14.358272 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:20:27 crc kubenswrapper[4906]: I1123 08:20:27.356852 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:20:27 crc kubenswrapper[4906]: E1123 08:20:27.358125 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:20:40 crc kubenswrapper[4906]: I1123 08:20:40.356292 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:20:40 crc kubenswrapper[4906]: E1123 08:20:40.357301 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:20:54 crc kubenswrapper[4906]: I1123 08:20:54.356508 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:20:54 crc kubenswrapper[4906]: E1123 08:20:54.357584 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:21:08 crc kubenswrapper[4906]: I1123 08:21:08.356820 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:21:08 crc kubenswrapper[4906]: E1123 08:21:08.357884 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:21:22 crc kubenswrapper[4906]: I1123 08:21:22.356892 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:21:22 crc kubenswrapper[4906]: E1123 08:21:22.357873 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:21:35 crc kubenswrapper[4906]: I1123 08:21:35.364472 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:21:35 crc kubenswrapper[4906]: E1123 08:21:35.365926 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:21:49 crc kubenswrapper[4906]: I1123 08:21:49.356769 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:21:49 crc kubenswrapper[4906]: E1123 08:21:49.373797 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:22:01 crc kubenswrapper[4906]: I1123 08:22:01.357488 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:22:01 crc kubenswrapper[4906]: E1123 08:22:01.359150 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.630199 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-brlmr"] Nov 23 08:22:08 crc kubenswrapper[4906]: E1123 08:22:08.631362 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="registry-server" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.631390 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="registry-server" Nov 23 08:22:08 crc kubenswrapper[4906]: E1123 08:22:08.631421 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="extract-utilities" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.631434 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="extract-utilities" Nov 23 08:22:08 crc kubenswrapper[4906]: E1123 08:22:08.631495 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="extract-content" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.631511 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="extract-content" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.631821 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdd5d275-a1c6-45af-832b-d511a444b3b8" containerName="registry-server" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.633886 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.639731 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brlmr"] Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.713472 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-utilities\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.713715 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-catalog-content\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.713792 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpzsz\" (UniqueName: \"kubernetes.io/projected/f936a245-2ccf-4c98-b2e3-72bbfa261cec-kube-api-access-lpzsz\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.814505 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-catalog-content\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.814571 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpzsz\" (UniqueName: \"kubernetes.io/projected/f936a245-2ccf-4c98-b2e3-72bbfa261cec-kube-api-access-lpzsz\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.814618 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-utilities\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.815033 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-catalog-content\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.815061 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-utilities\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.836370 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpzsz\" (UniqueName: \"kubernetes.io/projected/f936a245-2ccf-4c98-b2e3-72bbfa261cec-kube-api-access-lpzsz\") pod \"redhat-marketplace-brlmr\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:08 crc kubenswrapper[4906]: I1123 08:22:08.953250 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:09 crc kubenswrapper[4906]: I1123 08:22:09.444763 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brlmr"] Nov 23 08:22:09 crc kubenswrapper[4906]: I1123 08:22:09.827603 4906 generic.go:334] "Generic (PLEG): container finished" podID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerID="7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8" exitCode=0 Nov 23 08:22:09 crc kubenswrapper[4906]: I1123 08:22:09.827665 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brlmr" event={"ID":"f936a245-2ccf-4c98-b2e3-72bbfa261cec","Type":"ContainerDied","Data":"7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8"} Nov 23 08:22:09 crc kubenswrapper[4906]: I1123 08:22:09.827729 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brlmr" event={"ID":"f936a245-2ccf-4c98-b2e3-72bbfa261cec","Type":"ContainerStarted","Data":"f284732df0eb4e735614d9a3b19d6057d27a6276356dfa3a29e84d796bd6013e"} Nov 23 08:22:09 crc kubenswrapper[4906]: I1123 08:22:09.830007 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:22:10 crc kubenswrapper[4906]: I1123 08:22:10.847835 4906 generic.go:334] "Generic (PLEG): container finished" podID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerID="e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e" exitCode=0 Nov 23 08:22:10 crc kubenswrapper[4906]: I1123 08:22:10.848426 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brlmr" event={"ID":"f936a245-2ccf-4c98-b2e3-72bbfa261cec","Type":"ContainerDied","Data":"e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e"} Nov 23 08:22:11 crc kubenswrapper[4906]: I1123 08:22:11.861850 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brlmr" event={"ID":"f936a245-2ccf-4c98-b2e3-72bbfa261cec","Type":"ContainerStarted","Data":"95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2"} Nov 23 08:22:11 crc kubenswrapper[4906]: I1123 08:22:11.888598 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-brlmr" podStartSLOduration=2.424093757 podStartE2EDuration="3.88857789s" podCreationTimestamp="2025-11-23 08:22:08 +0000 UTC" firstStartedPulling="2025-11-23 08:22:09.829427682 +0000 UTC m=+5545.342819025" lastFinishedPulling="2025-11-23 08:22:11.293911815 +0000 UTC m=+5546.807303158" observedRunningTime="2025-11-23 08:22:11.888104767 +0000 UTC m=+5547.401496080" watchObservedRunningTime="2025-11-23 08:22:11.88857789 +0000 UTC m=+5547.401969203" Nov 23 08:22:14 crc kubenswrapper[4906]: I1123 08:22:14.356914 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:22:14 crc kubenswrapper[4906]: E1123 08:22:14.357455 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:22:18 crc kubenswrapper[4906]: I1123 08:22:18.954194 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:18 crc kubenswrapper[4906]: I1123 08:22:18.955279 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:19 crc kubenswrapper[4906]: I1123 08:22:19.017974 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:20 crc kubenswrapper[4906]: I1123 08:22:20.011431 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:20 crc kubenswrapper[4906]: I1123 08:22:20.091664 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-brlmr"] Nov 23 08:22:21 crc kubenswrapper[4906]: I1123 08:22:21.957947 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-brlmr" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="registry-server" containerID="cri-o://95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2" gracePeriod=2 Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.396331 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.481479 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-catalog-content\") pod \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.481540 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-utilities\") pod \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.481609 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpzsz\" (UniqueName: \"kubernetes.io/projected/f936a245-2ccf-4c98-b2e3-72bbfa261cec-kube-api-access-lpzsz\") pod \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\" (UID: \"f936a245-2ccf-4c98-b2e3-72bbfa261cec\") " Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.482675 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-utilities" (OuterVolumeSpecName: "utilities") pod "f936a245-2ccf-4c98-b2e3-72bbfa261cec" (UID: "f936a245-2ccf-4c98-b2e3-72bbfa261cec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.488943 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f936a245-2ccf-4c98-b2e3-72bbfa261cec-kube-api-access-lpzsz" (OuterVolumeSpecName: "kube-api-access-lpzsz") pod "f936a245-2ccf-4c98-b2e3-72bbfa261cec" (UID: "f936a245-2ccf-4c98-b2e3-72bbfa261cec"). InnerVolumeSpecName "kube-api-access-lpzsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.509074 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f936a245-2ccf-4c98-b2e3-72bbfa261cec" (UID: "f936a245-2ccf-4c98-b2e3-72bbfa261cec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.583543 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.583599 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f936a245-2ccf-4c98-b2e3-72bbfa261cec-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.583622 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpzsz\" (UniqueName: \"kubernetes.io/projected/f936a245-2ccf-4c98-b2e3-72bbfa261cec-kube-api-access-lpzsz\") on node \"crc\" DevicePath \"\"" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.976494 4906 generic.go:334] "Generic (PLEG): container finished" podID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerID="95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2" exitCode=0 Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.976550 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brlmr" event={"ID":"f936a245-2ccf-4c98-b2e3-72bbfa261cec","Type":"ContainerDied","Data":"95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2"} Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.976583 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brlmr" event={"ID":"f936a245-2ccf-4c98-b2e3-72bbfa261cec","Type":"ContainerDied","Data":"f284732df0eb4e735614d9a3b19d6057d27a6276356dfa3a29e84d796bd6013e"} Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.976584 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brlmr" Nov 23 08:22:22 crc kubenswrapper[4906]: I1123 08:22:22.976603 4906 scope.go:117] "RemoveContainer" containerID="95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.007621 4906 scope.go:117] "RemoveContainer" containerID="e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.026711 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-brlmr"] Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.039960 4906 scope.go:117] "RemoveContainer" containerID="7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.046052 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-brlmr"] Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.068719 4906 scope.go:117] "RemoveContainer" containerID="95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2" Nov 23 08:22:23 crc kubenswrapper[4906]: E1123 08:22:23.069654 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2\": container with ID starting with 95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2 not found: ID does not exist" containerID="95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.069718 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2"} err="failed to get container status \"95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2\": rpc error: code = NotFound desc = could not find container \"95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2\": container with ID starting with 95f26d525dd8920fa76eada87e0e322a40ddde8cdc6e4d1784ce6239b81515a2 not found: ID does not exist" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.069753 4906 scope.go:117] "RemoveContainer" containerID="e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e" Nov 23 08:22:23 crc kubenswrapper[4906]: E1123 08:22:23.070158 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e\": container with ID starting with e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e not found: ID does not exist" containerID="e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.070193 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e"} err="failed to get container status \"e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e\": rpc error: code = NotFound desc = could not find container \"e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e\": container with ID starting with e9c5bd0b94c3f868f6d12a1e275a1a884c3149e6199849405d318fcc03868f0e not found: ID does not exist" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.070219 4906 scope.go:117] "RemoveContainer" containerID="7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8" Nov 23 08:22:23 crc kubenswrapper[4906]: E1123 08:22:23.070519 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8\": container with ID starting with 7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8 not found: ID does not exist" containerID="7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.070551 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8"} err="failed to get container status \"7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8\": rpc error: code = NotFound desc = could not find container \"7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8\": container with ID starting with 7a1f041a8539265bab77e728746ce51813a7172b8e79786377afa0cdb5a202a8 not found: ID does not exist" Nov 23 08:22:23 crc kubenswrapper[4906]: I1123 08:22:23.365931 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" path="/var/lib/kubelet/pods/f936a245-2ccf-4c98-b2e3-72bbfa261cec/volumes" Nov 23 08:22:29 crc kubenswrapper[4906]: I1123 08:22:29.356977 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:22:29 crc kubenswrapper[4906]: E1123 08:22:29.358260 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:22:42 crc kubenswrapper[4906]: I1123 08:22:42.357013 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:22:42 crc kubenswrapper[4906]: E1123 08:22:42.358961 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:22:53 crc kubenswrapper[4906]: I1123 08:22:53.358206 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:22:54 crc kubenswrapper[4906]: I1123 08:22:54.237099 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"fb81892046b7234951e40cb90aa3bb89ca2103e0578651cdee5537c2372537e6"} Nov 23 08:25:20 crc kubenswrapper[4906]: I1123 08:25:20.945734 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:25:20 crc kubenswrapper[4906]: I1123 08:25:20.946466 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:25:50 crc kubenswrapper[4906]: I1123 08:25:50.945630 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:25:50 crc kubenswrapper[4906]: I1123 08:25:50.946501 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:26:20 crc kubenswrapper[4906]: I1123 08:26:20.946015 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:26:20 crc kubenswrapper[4906]: I1123 08:26:20.946626 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:26:20 crc kubenswrapper[4906]: I1123 08:26:20.946722 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:26:20 crc kubenswrapper[4906]: I1123 08:26:20.947475 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb81892046b7234951e40cb90aa3bb89ca2103e0578651cdee5537c2372537e6"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:26:20 crc kubenswrapper[4906]: I1123 08:26:20.947571 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://fb81892046b7234951e40cb90aa3bb89ca2103e0578651cdee5537c2372537e6" gracePeriod=600 Nov 23 08:26:21 crc kubenswrapper[4906]: I1123 08:26:21.685074 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="fb81892046b7234951e40cb90aa3bb89ca2103e0578651cdee5537c2372537e6" exitCode=0 Nov 23 08:26:21 crc kubenswrapper[4906]: I1123 08:26:21.685163 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"fb81892046b7234951e40cb90aa3bb89ca2103e0578651cdee5537c2372537e6"} Nov 23 08:26:21 crc kubenswrapper[4906]: I1123 08:26:21.685586 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809"} Nov 23 08:26:21 crc kubenswrapper[4906]: I1123 08:26:21.685640 4906 scope.go:117] "RemoveContainer" containerID="8163b161b207002ae8d458001bdd9744077c167b7d778dbbecf52354e7b1a3d6" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.266114 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-stf55"] Nov 23 08:27:26 crc kubenswrapper[4906]: E1123 08:27:26.266878 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="extract-content" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.266901 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="extract-content" Nov 23 08:27:26 crc kubenswrapper[4906]: E1123 08:27:26.266931 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="registry-server" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.266944 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="registry-server" Nov 23 08:27:26 crc kubenswrapper[4906]: E1123 08:27:26.266996 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="extract-utilities" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.267010 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="extract-utilities" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.267284 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="f936a245-2ccf-4c98-b2e3-72bbfa261cec" containerName="registry-server" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.275734 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.323940 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stf55"] Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.399800 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-utilities\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.399869 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-catalog-content\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.399952 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksdpn\" (UniqueName: \"kubernetes.io/projected/8ac654d4-2415-4f59-893f-f6523751538d-kube-api-access-ksdpn\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.501618 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-catalog-content\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.502038 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksdpn\" (UniqueName: \"kubernetes.io/projected/8ac654d4-2415-4f59-893f-f6523751538d-kube-api-access-ksdpn\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.502216 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-utilities\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.502697 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-catalog-content\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.502807 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-utilities\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.522435 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksdpn\" (UniqueName: \"kubernetes.io/projected/8ac654d4-2415-4f59-893f-f6523751538d-kube-api-access-ksdpn\") pod \"community-operators-stf55\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:26 crc kubenswrapper[4906]: I1123 08:27:26.639229 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:27 crc kubenswrapper[4906]: I1123 08:27:27.150356 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stf55"] Nov 23 08:27:27 crc kubenswrapper[4906]: I1123 08:27:27.323754 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stf55" event={"ID":"8ac654d4-2415-4f59-893f-f6523751538d","Type":"ContainerStarted","Data":"ec2b90ed8f3742482b80d7a0f19af5165e679de763a33a009ba1bfdc74cb484a"} Nov 23 08:27:28 crc kubenswrapper[4906]: I1123 08:27:28.333167 4906 generic.go:334] "Generic (PLEG): container finished" podID="8ac654d4-2415-4f59-893f-f6523751538d" containerID="1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0" exitCode=0 Nov 23 08:27:28 crc kubenswrapper[4906]: I1123 08:27:28.333217 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stf55" event={"ID":"8ac654d4-2415-4f59-893f-f6523751538d","Type":"ContainerDied","Data":"1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0"} Nov 23 08:27:28 crc kubenswrapper[4906]: I1123 08:27:28.335275 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:27:29 crc kubenswrapper[4906]: I1123 08:27:29.343901 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stf55" event={"ID":"8ac654d4-2415-4f59-893f-f6523751538d","Type":"ContainerStarted","Data":"7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685"} Nov 23 08:27:30 crc kubenswrapper[4906]: I1123 08:27:30.356655 4906 generic.go:334] "Generic (PLEG): container finished" podID="8ac654d4-2415-4f59-893f-f6523751538d" containerID="7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685" exitCode=0 Nov 23 08:27:30 crc kubenswrapper[4906]: I1123 08:27:30.356709 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stf55" event={"ID":"8ac654d4-2415-4f59-893f-f6523751538d","Type":"ContainerDied","Data":"7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685"} Nov 23 08:27:31 crc kubenswrapper[4906]: I1123 08:27:31.385701 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stf55" event={"ID":"8ac654d4-2415-4f59-893f-f6523751538d","Type":"ContainerStarted","Data":"954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a"} Nov 23 08:27:31 crc kubenswrapper[4906]: I1123 08:27:31.418664 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-stf55" podStartSLOduration=2.997706659 podStartE2EDuration="5.41864046s" podCreationTimestamp="2025-11-23 08:27:26 +0000 UTC" firstStartedPulling="2025-11-23 08:27:28.334967128 +0000 UTC m=+5863.848358441" lastFinishedPulling="2025-11-23 08:27:30.755900939 +0000 UTC m=+5866.269292242" observedRunningTime="2025-11-23 08:27:31.412033667 +0000 UTC m=+5866.925425000" watchObservedRunningTime="2025-11-23 08:27:31.41864046 +0000 UTC m=+5866.932031773" Nov 23 08:27:36 crc kubenswrapper[4906]: I1123 08:27:36.639548 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:36 crc kubenswrapper[4906]: I1123 08:27:36.640574 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:36 crc kubenswrapper[4906]: I1123 08:27:36.722959 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:37 crc kubenswrapper[4906]: I1123 08:27:37.507710 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:37 crc kubenswrapper[4906]: I1123 08:27:37.586210 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-stf55"] Nov 23 08:27:39 crc kubenswrapper[4906]: I1123 08:27:39.468211 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-stf55" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="registry-server" containerID="cri-o://954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a" gracePeriod=2 Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.037874 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.132621 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-catalog-content\") pod \"8ac654d4-2415-4f59-893f-f6523751538d\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.132857 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-utilities\") pod \"8ac654d4-2415-4f59-893f-f6523751538d\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.134216 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-utilities" (OuterVolumeSpecName: "utilities") pod "8ac654d4-2415-4f59-893f-f6523751538d" (UID: "8ac654d4-2415-4f59-893f-f6523751538d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.134445 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksdpn\" (UniqueName: \"kubernetes.io/projected/8ac654d4-2415-4f59-893f-f6523751538d-kube-api-access-ksdpn\") pod \"8ac654d4-2415-4f59-893f-f6523751538d\" (UID: \"8ac654d4-2415-4f59-893f-f6523751538d\") " Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.136430 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.142897 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ac654d4-2415-4f59-893f-f6523751538d-kube-api-access-ksdpn" (OuterVolumeSpecName: "kube-api-access-ksdpn") pod "8ac654d4-2415-4f59-893f-f6523751538d" (UID: "8ac654d4-2415-4f59-893f-f6523751538d"). InnerVolumeSpecName "kube-api-access-ksdpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.228299 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ac654d4-2415-4f59-893f-f6523751538d" (UID: "8ac654d4-2415-4f59-893f-f6523751538d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.238561 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ac654d4-2415-4f59-893f-f6523751538d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.238608 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksdpn\" (UniqueName: \"kubernetes.io/projected/8ac654d4-2415-4f59-893f-f6523751538d-kube-api-access-ksdpn\") on node \"crc\" DevicePath \"\"" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.482379 4906 generic.go:334] "Generic (PLEG): container finished" podID="8ac654d4-2415-4f59-893f-f6523751538d" containerID="954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a" exitCode=0 Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.482447 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stf55" event={"ID":"8ac654d4-2415-4f59-893f-f6523751538d","Type":"ContainerDied","Data":"954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a"} Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.482595 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stf55" event={"ID":"8ac654d4-2415-4f59-893f-f6523751538d","Type":"ContainerDied","Data":"ec2b90ed8f3742482b80d7a0f19af5165e679de763a33a009ba1bfdc74cb484a"} Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.482504 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stf55" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.482651 4906 scope.go:117] "RemoveContainer" containerID="954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.520958 4906 scope.go:117] "RemoveContainer" containerID="7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.570743 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-stf55"] Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.570889 4906 scope.go:117] "RemoveContainer" containerID="1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.589415 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-stf55"] Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.593755 4906 scope.go:117] "RemoveContainer" containerID="954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a" Nov 23 08:27:40 crc kubenswrapper[4906]: E1123 08:27:40.594346 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a\": container with ID starting with 954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a not found: ID does not exist" containerID="954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.594404 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a"} err="failed to get container status \"954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a\": rpc error: code = NotFound desc = could not find container \"954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a\": container with ID starting with 954624db93fc4c32b7a8330ca1a3484c08ab4259c0dd2e71b451066f6361195a not found: ID does not exist" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.594443 4906 scope.go:117] "RemoveContainer" containerID="7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685" Nov 23 08:27:40 crc kubenswrapper[4906]: E1123 08:27:40.595376 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685\": container with ID starting with 7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685 not found: ID does not exist" containerID="7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.595424 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685"} err="failed to get container status \"7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685\": rpc error: code = NotFound desc = could not find container \"7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685\": container with ID starting with 7352229f0715eec6365e0eacd5fb16e973def195a244b94a4ee25657efedb685 not found: ID does not exist" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.595455 4906 scope.go:117] "RemoveContainer" containerID="1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0" Nov 23 08:27:40 crc kubenswrapper[4906]: E1123 08:27:40.596035 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0\": container with ID starting with 1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0 not found: ID does not exist" containerID="1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0" Nov 23 08:27:40 crc kubenswrapper[4906]: I1123 08:27:40.596085 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0"} err="failed to get container status \"1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0\": rpc error: code = NotFound desc = could not find container \"1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0\": container with ID starting with 1d7eee25fab3ae6db51945e17ef6cc8d1a76cd75d2ef6ac0273bfda91a42c0f0 not found: ID does not exist" Nov 23 08:27:41 crc kubenswrapper[4906]: I1123 08:27:41.372151 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ac654d4-2415-4f59-893f-f6523751538d" path="/var/lib/kubelet/pods/8ac654d4-2415-4f59-893f-f6523751538d/volumes" Nov 23 08:28:50 crc kubenswrapper[4906]: I1123 08:28:50.946330 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:28:50 crc kubenswrapper[4906]: I1123 08:28:50.947109 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:29:20 crc kubenswrapper[4906]: I1123 08:29:20.945457 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:29:20 crc kubenswrapper[4906]: I1123 08:29:20.946248 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:29:50 crc kubenswrapper[4906]: I1123 08:29:50.945928 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:29:50 crc kubenswrapper[4906]: I1123 08:29:50.946773 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:29:50 crc kubenswrapper[4906]: I1123 08:29:50.946838 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:29:50 crc kubenswrapper[4906]: I1123 08:29:50.947779 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:29:50 crc kubenswrapper[4906]: I1123 08:29:50.947868 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" gracePeriod=600 Nov 23 08:29:51 crc kubenswrapper[4906]: E1123 08:29:51.081057 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:29:51 crc kubenswrapper[4906]: I1123 08:29:51.702636 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" exitCode=0 Nov 23 08:29:51 crc kubenswrapper[4906]: I1123 08:29:51.702729 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809"} Nov 23 08:29:51 crc kubenswrapper[4906]: I1123 08:29:51.703141 4906 scope.go:117] "RemoveContainer" containerID="fb81892046b7234951e40cb90aa3bb89ca2103e0578651cdee5537c2372537e6" Nov 23 08:29:51 crc kubenswrapper[4906]: I1123 08:29:51.704120 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:29:51 crc kubenswrapper[4906]: E1123 08:29:51.704516 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.148879 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt"] Nov 23 08:30:00 crc kubenswrapper[4906]: E1123 08:30:00.149632 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="registry-server" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.149646 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="registry-server" Nov 23 08:30:00 crc kubenswrapper[4906]: E1123 08:30:00.149663 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="extract-utilities" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.149669 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="extract-utilities" Nov 23 08:30:00 crc kubenswrapper[4906]: E1123 08:30:00.149697 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="extract-content" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.149704 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="extract-content" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.149848 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ac654d4-2415-4f59-893f-f6523751538d" containerName="registry-server" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.150312 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.154518 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.154638 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.164891 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt"] Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.248138 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfl2x\" (UniqueName: \"kubernetes.io/projected/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-kube-api-access-vfl2x\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.248203 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-config-volume\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.248404 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-secret-volume\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.350070 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-secret-volume\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.350165 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfl2x\" (UniqueName: \"kubernetes.io/projected/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-kube-api-access-vfl2x\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.350197 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-config-volume\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.351278 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-config-volume\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.356761 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-secret-volume\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.365968 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfl2x\" (UniqueName: \"kubernetes.io/projected/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-kube-api-access-vfl2x\") pod \"collect-profiles-29398110-mscxt\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.484921 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.707790 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt"] Nov 23 08:30:00 crc kubenswrapper[4906]: I1123 08:30:00.786478 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" event={"ID":"bd83ec29-3282-4a78-a3ed-5ecae24a1c03","Type":"ContainerStarted","Data":"cc322d044ce4b20bfbde7d031daea6a0bb7df5029f4cdee66f9559c1321b6150"} Nov 23 08:30:01 crc kubenswrapper[4906]: I1123 08:30:01.793724 4906 generic.go:334] "Generic (PLEG): container finished" podID="bd83ec29-3282-4a78-a3ed-5ecae24a1c03" containerID="5a257131a9457be7fc5f8a907eac9c7aeef2793982320f3a77456f4b13738d68" exitCode=0 Nov 23 08:30:01 crc kubenswrapper[4906]: I1123 08:30:01.793885 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" event={"ID":"bd83ec29-3282-4a78-a3ed-5ecae24a1c03","Type":"ContainerDied","Data":"5a257131a9457be7fc5f8a907eac9c7aeef2793982320f3a77456f4b13738d68"} Nov 23 08:30:02 crc kubenswrapper[4906]: I1123 08:30:02.356485 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:30:02 crc kubenswrapper[4906]: E1123 08:30:02.356941 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.180431 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.314057 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfl2x\" (UniqueName: \"kubernetes.io/projected/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-kube-api-access-vfl2x\") pod \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.314111 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-secret-volume\") pod \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.314177 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-config-volume\") pod \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\" (UID: \"bd83ec29-3282-4a78-a3ed-5ecae24a1c03\") " Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.316242 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-config-volume" (OuterVolumeSpecName: "config-volume") pod "bd83ec29-3282-4a78-a3ed-5ecae24a1c03" (UID: "bd83ec29-3282-4a78-a3ed-5ecae24a1c03"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.320644 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bd83ec29-3282-4a78-a3ed-5ecae24a1c03" (UID: "bd83ec29-3282-4a78-a3ed-5ecae24a1c03"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.328714 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-kube-api-access-vfl2x" (OuterVolumeSpecName: "kube-api-access-vfl2x") pod "bd83ec29-3282-4a78-a3ed-5ecae24a1c03" (UID: "bd83ec29-3282-4a78-a3ed-5ecae24a1c03"). InnerVolumeSpecName "kube-api-access-vfl2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.416382 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.416448 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfl2x\" (UniqueName: \"kubernetes.io/projected/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-kube-api-access-vfl2x\") on node \"crc\" DevicePath \"\"" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.416473 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd83ec29-3282-4a78-a3ed-5ecae24a1c03-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.817464 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" event={"ID":"bd83ec29-3282-4a78-a3ed-5ecae24a1c03","Type":"ContainerDied","Data":"cc322d044ce4b20bfbde7d031daea6a0bb7df5029f4cdee66f9559c1321b6150"} Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.817532 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc322d044ce4b20bfbde7d031daea6a0bb7df5029f4cdee66f9559c1321b6150" Nov 23 08:30:03 crc kubenswrapper[4906]: I1123 08:30:03.817579 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398110-mscxt" Nov 23 08:30:04 crc kubenswrapper[4906]: I1123 08:30:04.282515 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g"] Nov 23 08:30:04 crc kubenswrapper[4906]: I1123 08:30:04.289462 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398065-2ct6g"] Nov 23 08:30:05 crc kubenswrapper[4906]: I1123 08:30:05.373767 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f46478a6-465d-4e20-b56a-54904cc46116" path="/var/lib/kubelet/pods/f46478a6-465d-4e20-b56a-54904cc46116/volumes" Nov 23 08:30:14 crc kubenswrapper[4906]: I1123 08:30:14.356460 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:30:14 crc kubenswrapper[4906]: E1123 08:30:14.357544 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:30:27 crc kubenswrapper[4906]: I1123 08:30:27.356403 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:30:27 crc kubenswrapper[4906]: E1123 08:30:27.357137 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:30:38 crc kubenswrapper[4906]: I1123 08:30:38.356996 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:30:38 crc kubenswrapper[4906]: E1123 08:30:38.358176 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:30:53 crc kubenswrapper[4906]: I1123 08:30:53.356855 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:30:53 crc kubenswrapper[4906]: E1123 08:30:53.358756 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.152896 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c4sd8"] Nov 23 08:30:56 crc kubenswrapper[4906]: E1123 08:30:56.153540 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd83ec29-3282-4a78-a3ed-5ecae24a1c03" containerName="collect-profiles" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.153558 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd83ec29-3282-4a78-a3ed-5ecae24a1c03" containerName="collect-profiles" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.153802 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd83ec29-3282-4a78-a3ed-5ecae24a1c03" containerName="collect-profiles" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.155101 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.176631 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4sd8"] Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.270858 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-utilities\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.270908 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-catalog-content\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.270933 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gt86\" (UniqueName: \"kubernetes.io/projected/065a78b7-09c8-4e4c-8fc7-4e76c322f821-kube-api-access-9gt86\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.371948 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-utilities\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.371992 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-catalog-content\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.372014 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gt86\" (UniqueName: \"kubernetes.io/projected/065a78b7-09c8-4e4c-8fc7-4e76c322f821-kube-api-access-9gt86\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.372427 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-catalog-content\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.372482 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-utilities\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.391058 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gt86\" (UniqueName: \"kubernetes.io/projected/065a78b7-09c8-4e4c-8fc7-4e76c322f821-kube-api-access-9gt86\") pod \"certified-operators-c4sd8\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.473824 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:30:56 crc kubenswrapper[4906]: I1123 08:30:56.922823 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4sd8"] Nov 23 08:30:57 crc kubenswrapper[4906]: I1123 08:30:57.259985 4906 generic.go:334] "Generic (PLEG): container finished" podID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerID="ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553" exitCode=0 Nov 23 08:30:57 crc kubenswrapper[4906]: I1123 08:30:57.260027 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4sd8" event={"ID":"065a78b7-09c8-4e4c-8fc7-4e76c322f821","Type":"ContainerDied","Data":"ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553"} Nov 23 08:30:57 crc kubenswrapper[4906]: I1123 08:30:57.260053 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4sd8" event={"ID":"065a78b7-09c8-4e4c-8fc7-4e76c322f821","Type":"ContainerStarted","Data":"f61e856d6fa5131cebab30b9859f102f68860fb3facee17b6b26040f844d4f3a"} Nov 23 08:30:57 crc kubenswrapper[4906]: I1123 08:30:57.944462 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h765h"] Nov 23 08:30:57 crc kubenswrapper[4906]: I1123 08:30:57.946172 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:57 crc kubenswrapper[4906]: I1123 08:30:57.962248 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h765h"] Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.002499 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-utilities\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.002645 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-catalog-content\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.002738 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv2lf\" (UniqueName: \"kubernetes.io/projected/03659852-f619-478b-a117-820d211b623a-kube-api-access-wv2lf\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.103970 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-catalog-content\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.104053 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv2lf\" (UniqueName: \"kubernetes.io/projected/03659852-f619-478b-a117-820d211b623a-kube-api-access-wv2lf\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.104125 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-utilities\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.104446 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-catalog-content\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.104582 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-utilities\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.123618 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv2lf\" (UniqueName: \"kubernetes.io/projected/03659852-f619-478b-a117-820d211b623a-kube-api-access-wv2lf\") pod \"redhat-operators-h765h\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.269155 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4sd8" event={"ID":"065a78b7-09c8-4e4c-8fc7-4e76c322f821","Type":"ContainerStarted","Data":"e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664"} Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.297706 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.678608 4906 scope.go:117] "RemoveContainer" containerID="b2f9731718c7c464e81bb2a2b0dfd45be0c1218fa1458f99d657e33d28cf414e" Nov 23 08:30:58 crc kubenswrapper[4906]: I1123 08:30:58.749318 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h765h"] Nov 23 08:30:58 crc kubenswrapper[4906]: W1123 08:30:58.752696 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03659852_f619_478b_a117_820d211b623a.slice/crio-930a5a1be8f11033c48e7e85976f71a4ef651e3cefafa426a0680fb59b9d3cdc WatchSource:0}: Error finding container 930a5a1be8f11033c48e7e85976f71a4ef651e3cefafa426a0680fb59b9d3cdc: Status 404 returned error can't find the container with id 930a5a1be8f11033c48e7e85976f71a4ef651e3cefafa426a0680fb59b9d3cdc Nov 23 08:30:59 crc kubenswrapper[4906]: I1123 08:30:59.279382 4906 generic.go:334] "Generic (PLEG): container finished" podID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerID="e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664" exitCode=0 Nov 23 08:30:59 crc kubenswrapper[4906]: I1123 08:30:59.279712 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4sd8" event={"ID":"065a78b7-09c8-4e4c-8fc7-4e76c322f821","Type":"ContainerDied","Data":"e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664"} Nov 23 08:30:59 crc kubenswrapper[4906]: I1123 08:30:59.281811 4906 generic.go:334] "Generic (PLEG): container finished" podID="03659852-f619-478b-a117-820d211b623a" containerID="618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25" exitCode=0 Nov 23 08:30:59 crc kubenswrapper[4906]: I1123 08:30:59.281861 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h765h" event={"ID":"03659852-f619-478b-a117-820d211b623a","Type":"ContainerDied","Data":"618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25"} Nov 23 08:30:59 crc kubenswrapper[4906]: I1123 08:30:59.281896 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h765h" event={"ID":"03659852-f619-478b-a117-820d211b623a","Type":"ContainerStarted","Data":"930a5a1be8f11033c48e7e85976f71a4ef651e3cefafa426a0680fb59b9d3cdc"} Nov 23 08:31:00 crc kubenswrapper[4906]: I1123 08:31:00.296617 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h765h" event={"ID":"03659852-f619-478b-a117-820d211b623a","Type":"ContainerStarted","Data":"8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800"} Nov 23 08:31:00 crc kubenswrapper[4906]: I1123 08:31:00.301163 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4sd8" event={"ID":"065a78b7-09c8-4e4c-8fc7-4e76c322f821","Type":"ContainerStarted","Data":"353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8"} Nov 23 08:31:00 crc kubenswrapper[4906]: I1123 08:31:00.350952 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c4sd8" podStartSLOduration=1.944887971 podStartE2EDuration="4.350934213s" podCreationTimestamp="2025-11-23 08:30:56 +0000 UTC" firstStartedPulling="2025-11-23 08:30:57.261855736 +0000 UTC m=+6072.775247039" lastFinishedPulling="2025-11-23 08:30:59.667901988 +0000 UTC m=+6075.181293281" observedRunningTime="2025-11-23 08:31:00.343641091 +0000 UTC m=+6075.857032404" watchObservedRunningTime="2025-11-23 08:31:00.350934213 +0000 UTC m=+6075.864325526" Nov 23 08:31:01 crc kubenswrapper[4906]: I1123 08:31:01.311915 4906 generic.go:334] "Generic (PLEG): container finished" podID="03659852-f619-478b-a117-820d211b623a" containerID="8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800" exitCode=0 Nov 23 08:31:01 crc kubenswrapper[4906]: I1123 08:31:01.311978 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h765h" event={"ID":"03659852-f619-478b-a117-820d211b623a","Type":"ContainerDied","Data":"8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800"} Nov 23 08:31:02 crc kubenswrapper[4906]: I1123 08:31:02.324672 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h765h" event={"ID":"03659852-f619-478b-a117-820d211b623a","Type":"ContainerStarted","Data":"67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677"} Nov 23 08:31:02 crc kubenswrapper[4906]: I1123 08:31:02.350291 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h765h" podStartSLOduration=2.876475602 podStartE2EDuration="5.350250282s" podCreationTimestamp="2025-11-23 08:30:57 +0000 UTC" firstStartedPulling="2025-11-23 08:30:59.28328553 +0000 UTC m=+6074.796676853" lastFinishedPulling="2025-11-23 08:31:01.75706023 +0000 UTC m=+6077.270451533" observedRunningTime="2025-11-23 08:31:02.348291831 +0000 UTC m=+6077.861683144" watchObservedRunningTime="2025-11-23 08:31:02.350250282 +0000 UTC m=+6077.863641605" Nov 23 08:31:06 crc kubenswrapper[4906]: I1123 08:31:06.474219 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:31:06 crc kubenswrapper[4906]: I1123 08:31:06.474560 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:31:06 crc kubenswrapper[4906]: I1123 08:31:06.529768 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:31:07 crc kubenswrapper[4906]: I1123 08:31:07.415423 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:31:07 crc kubenswrapper[4906]: I1123 08:31:07.496045 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4sd8"] Nov 23 08:31:08 crc kubenswrapper[4906]: I1123 08:31:08.297846 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:31:08 crc kubenswrapper[4906]: I1123 08:31:08.298275 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:31:08 crc kubenswrapper[4906]: I1123 08:31:08.349478 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:31:08 crc kubenswrapper[4906]: I1123 08:31:08.358099 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:31:08 crc kubenswrapper[4906]: E1123 08:31:08.359028 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:31:08 crc kubenswrapper[4906]: I1123 08:31:08.419728 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.178919 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h765h"] Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.384340 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c4sd8" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="registry-server" containerID="cri-o://353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8" gracePeriod=2 Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.831590 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.936458 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gt86\" (UniqueName: \"kubernetes.io/projected/065a78b7-09c8-4e4c-8fc7-4e76c322f821-kube-api-access-9gt86\") pod \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.936504 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-catalog-content\") pod \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.936558 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-utilities\") pod \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\" (UID: \"065a78b7-09c8-4e4c-8fc7-4e76c322f821\") " Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.937627 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-utilities" (OuterVolumeSpecName: "utilities") pod "065a78b7-09c8-4e4c-8fc7-4e76c322f821" (UID: "065a78b7-09c8-4e4c-8fc7-4e76c322f821"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.942499 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065a78b7-09c8-4e4c-8fc7-4e76c322f821-kube-api-access-9gt86" (OuterVolumeSpecName: "kube-api-access-9gt86") pod "065a78b7-09c8-4e4c-8fc7-4e76c322f821" (UID: "065a78b7-09c8-4e4c-8fc7-4e76c322f821"). InnerVolumeSpecName "kube-api-access-9gt86". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:31:09 crc kubenswrapper[4906]: I1123 08:31:09.997500 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "065a78b7-09c8-4e4c-8fc7-4e76c322f821" (UID: "065a78b7-09c8-4e4c-8fc7-4e76c322f821"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.037954 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.037988 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gt86\" (UniqueName: \"kubernetes.io/projected/065a78b7-09c8-4e4c-8fc7-4e76c322f821-kube-api-access-9gt86\") on node \"crc\" DevicePath \"\"" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.037999 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a78b7-09c8-4e4c-8fc7-4e76c322f821-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.393019 4906 generic.go:334] "Generic (PLEG): container finished" podID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerID="353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8" exitCode=0 Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.394212 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4sd8" event={"ID":"065a78b7-09c8-4e4c-8fc7-4e76c322f821","Type":"ContainerDied","Data":"353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8"} Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.394260 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4sd8" event={"ID":"065a78b7-09c8-4e4c-8fc7-4e76c322f821","Type":"ContainerDied","Data":"f61e856d6fa5131cebab30b9859f102f68860fb3facee17b6b26040f844d4f3a"} Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.394309 4906 scope.go:117] "RemoveContainer" containerID="353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.394446 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4sd8" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.394713 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h765h" podUID="03659852-f619-478b-a117-820d211b623a" containerName="registry-server" containerID="cri-o://67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677" gracePeriod=2 Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.428420 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4sd8"] Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.431276 4906 scope.go:117] "RemoveContainer" containerID="e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.433299 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c4sd8"] Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.458396 4906 scope.go:117] "RemoveContainer" containerID="ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.545598 4906 scope.go:117] "RemoveContainer" containerID="353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8" Nov 23 08:31:10 crc kubenswrapper[4906]: E1123 08:31:10.549594 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8\": container with ID starting with 353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8 not found: ID does not exist" containerID="353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.549792 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8"} err="failed to get container status \"353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8\": rpc error: code = NotFound desc = could not find container \"353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8\": container with ID starting with 353df13381c165e505e42432815a060c1213d58ff8a502365e2cd571c9042ff8 not found: ID does not exist" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.549858 4906 scope.go:117] "RemoveContainer" containerID="e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664" Nov 23 08:31:10 crc kubenswrapper[4906]: E1123 08:31:10.550365 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664\": container with ID starting with e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664 not found: ID does not exist" containerID="e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.550413 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664"} err="failed to get container status \"e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664\": rpc error: code = NotFound desc = could not find container \"e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664\": container with ID starting with e5308f36a312196425d9d3e869e1b2eb51905b7aa06d597ce0040b18abaf9664 not found: ID does not exist" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.550441 4906 scope.go:117] "RemoveContainer" containerID="ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553" Nov 23 08:31:10 crc kubenswrapper[4906]: E1123 08:31:10.550992 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553\": container with ID starting with ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553 not found: ID does not exist" containerID="ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.551039 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553"} err="failed to get container status \"ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553\": rpc error: code = NotFound desc = could not find container \"ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553\": container with ID starting with ec53640d2553d0635075846db9d8d1d262dd8490bbac51b408c53705eb17b553 not found: ID does not exist" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.840231 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.972460 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv2lf\" (UniqueName: \"kubernetes.io/projected/03659852-f619-478b-a117-820d211b623a-kube-api-access-wv2lf\") pod \"03659852-f619-478b-a117-820d211b623a\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.972558 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-catalog-content\") pod \"03659852-f619-478b-a117-820d211b623a\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.972705 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-utilities\") pod \"03659852-f619-478b-a117-820d211b623a\" (UID: \"03659852-f619-478b-a117-820d211b623a\") " Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.973625 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-utilities" (OuterVolumeSpecName: "utilities") pod "03659852-f619-478b-a117-820d211b623a" (UID: "03659852-f619-478b-a117-820d211b623a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:31:10 crc kubenswrapper[4906]: I1123 08:31:10.979983 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03659852-f619-478b-a117-820d211b623a-kube-api-access-wv2lf" (OuterVolumeSpecName: "kube-api-access-wv2lf") pod "03659852-f619-478b-a117-820d211b623a" (UID: "03659852-f619-478b-a117-820d211b623a"). InnerVolumeSpecName "kube-api-access-wv2lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.067363 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03659852-f619-478b-a117-820d211b623a" (UID: "03659852-f619-478b-a117-820d211b623a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.074507 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv2lf\" (UniqueName: \"kubernetes.io/projected/03659852-f619-478b-a117-820d211b623a-kube-api-access-wv2lf\") on node \"crc\" DevicePath \"\"" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.074566 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.074584 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03659852-f619-478b-a117-820d211b623a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.370299 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" path="/var/lib/kubelet/pods/065a78b7-09c8-4e4c-8fc7-4e76c322f821/volumes" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.408733 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h765h" event={"ID":"03659852-f619-478b-a117-820d211b623a","Type":"ContainerDied","Data":"67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677"} Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.408846 4906 scope.go:117] "RemoveContainer" containerID="67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.408714 4906 generic.go:334] "Generic (PLEG): container finished" podID="03659852-f619-478b-a117-820d211b623a" containerID="67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677" exitCode=0 Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.408921 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h765h" event={"ID":"03659852-f619-478b-a117-820d211b623a","Type":"ContainerDied","Data":"930a5a1be8f11033c48e7e85976f71a4ef651e3cefafa426a0680fb59b9d3cdc"} Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.408765 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h765h" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.434786 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h765h"] Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.440540 4906 scope.go:117] "RemoveContainer" containerID="8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.441459 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h765h"] Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.459829 4906 scope.go:117] "RemoveContainer" containerID="618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.480431 4906 scope.go:117] "RemoveContainer" containerID="67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677" Nov 23 08:31:11 crc kubenswrapper[4906]: E1123 08:31:11.480944 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677\": container with ID starting with 67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677 not found: ID does not exist" containerID="67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.480974 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677"} err="failed to get container status \"67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677\": rpc error: code = NotFound desc = could not find container \"67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677\": container with ID starting with 67f03c8a145acd0994243ab1fb3378530793ca3d8e78df5dc7e12ba91c530677 not found: ID does not exist" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.480995 4906 scope.go:117] "RemoveContainer" containerID="8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800" Nov 23 08:31:11 crc kubenswrapper[4906]: E1123 08:31:11.481294 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800\": container with ID starting with 8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800 not found: ID does not exist" containerID="8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.481397 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800"} err="failed to get container status \"8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800\": rpc error: code = NotFound desc = could not find container \"8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800\": container with ID starting with 8c524a3de37c1fa91a749ee88aff6164e9c79aabc5324e442f7d8e24d3d83800 not found: ID does not exist" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.481473 4906 scope.go:117] "RemoveContainer" containerID="618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25" Nov 23 08:31:11 crc kubenswrapper[4906]: E1123 08:31:11.481907 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25\": container with ID starting with 618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25 not found: ID does not exist" containerID="618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25" Nov 23 08:31:11 crc kubenswrapper[4906]: I1123 08:31:11.481958 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25"} err="failed to get container status \"618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25\": rpc error: code = NotFound desc = could not find container \"618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25\": container with ID starting with 618203492a1745c7b6971d418a169ae459d80e0245d6617016b50337ce605f25 not found: ID does not exist" Nov 23 08:31:13 crc kubenswrapper[4906]: I1123 08:31:13.372848 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03659852-f619-478b-a117-820d211b623a" path="/var/lib/kubelet/pods/03659852-f619-478b-a117-820d211b623a/volumes" Nov 23 08:31:19 crc kubenswrapper[4906]: I1123 08:31:19.357166 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:31:19 crc kubenswrapper[4906]: E1123 08:31:19.358164 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:31:34 crc kubenswrapper[4906]: I1123 08:31:34.357002 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:31:34 crc kubenswrapper[4906]: E1123 08:31:34.358142 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:31:45 crc kubenswrapper[4906]: I1123 08:31:45.393363 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:31:45 crc kubenswrapper[4906]: E1123 08:31:45.394991 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:31:59 crc kubenswrapper[4906]: I1123 08:31:59.356236 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:31:59 crc kubenswrapper[4906]: E1123 08:31:59.357277 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:32:12 crc kubenswrapper[4906]: I1123 08:32:12.356630 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:32:12 crc kubenswrapper[4906]: E1123 08:32:12.357484 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:32:25 crc kubenswrapper[4906]: I1123 08:32:25.364291 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:32:25 crc kubenswrapper[4906]: E1123 08:32:25.365599 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:32:36 crc kubenswrapper[4906]: I1123 08:32:36.356588 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:32:36 crc kubenswrapper[4906]: E1123 08:32:36.357352 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:32:50 crc kubenswrapper[4906]: I1123 08:32:50.357793 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:32:50 crc kubenswrapper[4906]: E1123 08:32:50.358915 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.410311 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jxk2r"] Nov 23 08:32:53 crc kubenswrapper[4906]: E1123 08:32:53.411058 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03659852-f619-478b-a117-820d211b623a" containerName="extract-utilities" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411073 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03659852-f619-478b-a117-820d211b623a" containerName="extract-utilities" Nov 23 08:32:53 crc kubenswrapper[4906]: E1123 08:32:53.411089 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03659852-f619-478b-a117-820d211b623a" containerName="extract-content" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411098 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03659852-f619-478b-a117-820d211b623a" containerName="extract-content" Nov 23 08:32:53 crc kubenswrapper[4906]: E1123 08:32:53.411108 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03659852-f619-478b-a117-820d211b623a" containerName="registry-server" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411117 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03659852-f619-478b-a117-820d211b623a" containerName="registry-server" Nov 23 08:32:53 crc kubenswrapper[4906]: E1123 08:32:53.411131 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="extract-utilities" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411140 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="extract-utilities" Nov 23 08:32:53 crc kubenswrapper[4906]: E1123 08:32:53.411173 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="registry-server" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411180 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="registry-server" Nov 23 08:32:53 crc kubenswrapper[4906]: E1123 08:32:53.411195 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="extract-content" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411205 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="extract-content" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411390 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="03659852-f619-478b-a117-820d211b623a" containerName="registry-server" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.411414 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="065a78b7-09c8-4e4c-8fc7-4e76c322f821" containerName="registry-server" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.412899 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.433239 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxk2r"] Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.476659 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-catalog-content\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.476787 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-utilities\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.476994 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62z6m\" (UniqueName: \"kubernetes.io/projected/62b08d58-5a82-4492-93e1-eb0df006298c-kube-api-access-62z6m\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.577943 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62z6m\" (UniqueName: \"kubernetes.io/projected/62b08d58-5a82-4492-93e1-eb0df006298c-kube-api-access-62z6m\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.578070 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-catalog-content\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.578112 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-utilities\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.578869 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-catalog-content\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.578894 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-utilities\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.602842 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62z6m\" (UniqueName: \"kubernetes.io/projected/62b08d58-5a82-4492-93e1-eb0df006298c-kube-api-access-62z6m\") pod \"redhat-marketplace-jxk2r\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:53 crc kubenswrapper[4906]: I1123 08:32:53.741165 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:32:54 crc kubenswrapper[4906]: I1123 08:32:54.310006 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxk2r"] Nov 23 08:32:54 crc kubenswrapper[4906]: W1123 08:32:54.319175 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62b08d58_5a82_4492_93e1_eb0df006298c.slice/crio-2cf4775ce10678c3d3f047d5efca243d8f468465b642d2db0ea4a1a06b308dc1 WatchSource:0}: Error finding container 2cf4775ce10678c3d3f047d5efca243d8f468465b642d2db0ea4a1a06b308dc1: Status 404 returned error can't find the container with id 2cf4775ce10678c3d3f047d5efca243d8f468465b642d2db0ea4a1a06b308dc1 Nov 23 08:32:54 crc kubenswrapper[4906]: I1123 08:32:54.465130 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxk2r" event={"ID":"62b08d58-5a82-4492-93e1-eb0df006298c","Type":"ContainerStarted","Data":"2cf4775ce10678c3d3f047d5efca243d8f468465b642d2db0ea4a1a06b308dc1"} Nov 23 08:32:55 crc kubenswrapper[4906]: I1123 08:32:55.477114 4906 generic.go:334] "Generic (PLEG): container finished" podID="62b08d58-5a82-4492-93e1-eb0df006298c" containerID="070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f" exitCode=0 Nov 23 08:32:55 crc kubenswrapper[4906]: I1123 08:32:55.477441 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxk2r" event={"ID":"62b08d58-5a82-4492-93e1-eb0df006298c","Type":"ContainerDied","Data":"070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f"} Nov 23 08:32:55 crc kubenswrapper[4906]: I1123 08:32:55.479215 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:32:56 crc kubenswrapper[4906]: I1123 08:32:56.488236 4906 generic.go:334] "Generic (PLEG): container finished" podID="62b08d58-5a82-4492-93e1-eb0df006298c" containerID="d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54" exitCode=0 Nov 23 08:32:56 crc kubenswrapper[4906]: I1123 08:32:56.488334 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxk2r" event={"ID":"62b08d58-5a82-4492-93e1-eb0df006298c","Type":"ContainerDied","Data":"d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54"} Nov 23 08:32:57 crc kubenswrapper[4906]: I1123 08:32:57.503082 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxk2r" event={"ID":"62b08d58-5a82-4492-93e1-eb0df006298c","Type":"ContainerStarted","Data":"fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf"} Nov 23 08:32:57 crc kubenswrapper[4906]: I1123 08:32:57.524953 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jxk2r" podStartSLOduration=3.123038117 podStartE2EDuration="4.524917991s" podCreationTimestamp="2025-11-23 08:32:53 +0000 UTC" firstStartedPulling="2025-11-23 08:32:55.47895699 +0000 UTC m=+6190.992348293" lastFinishedPulling="2025-11-23 08:32:56.880836864 +0000 UTC m=+6192.394228167" observedRunningTime="2025-11-23 08:32:57.520422612 +0000 UTC m=+6193.033813965" watchObservedRunningTime="2025-11-23 08:32:57.524917991 +0000 UTC m=+6193.038309344" Nov 23 08:33:03 crc kubenswrapper[4906]: I1123 08:33:03.742032 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:33:03 crc kubenswrapper[4906]: I1123 08:33:03.742506 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:33:03 crc kubenswrapper[4906]: I1123 08:33:03.797812 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:33:04 crc kubenswrapper[4906]: I1123 08:33:04.356862 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:33:04 crc kubenswrapper[4906]: E1123 08:33:04.357168 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:33:04 crc kubenswrapper[4906]: I1123 08:33:04.636115 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:33:04 crc kubenswrapper[4906]: I1123 08:33:04.706267 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxk2r"] Nov 23 08:33:06 crc kubenswrapper[4906]: I1123 08:33:06.580218 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jxk2r" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="registry-server" containerID="cri-o://fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf" gracePeriod=2 Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.000067 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.086860 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62z6m\" (UniqueName: \"kubernetes.io/projected/62b08d58-5a82-4492-93e1-eb0df006298c-kube-api-access-62z6m\") pod \"62b08d58-5a82-4492-93e1-eb0df006298c\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.086916 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-utilities\") pod \"62b08d58-5a82-4492-93e1-eb0df006298c\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.086951 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-catalog-content\") pod \"62b08d58-5a82-4492-93e1-eb0df006298c\" (UID: \"62b08d58-5a82-4492-93e1-eb0df006298c\") " Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.088735 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-utilities" (OuterVolumeSpecName: "utilities") pod "62b08d58-5a82-4492-93e1-eb0df006298c" (UID: "62b08d58-5a82-4492-93e1-eb0df006298c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.099354 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b08d58-5a82-4492-93e1-eb0df006298c-kube-api-access-62z6m" (OuterVolumeSpecName: "kube-api-access-62z6m") pod "62b08d58-5a82-4492-93e1-eb0df006298c" (UID: "62b08d58-5a82-4492-93e1-eb0df006298c"). InnerVolumeSpecName "kube-api-access-62z6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.110753 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62b08d58-5a82-4492-93e1-eb0df006298c" (UID: "62b08d58-5a82-4492-93e1-eb0df006298c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.189283 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62z6m\" (UniqueName: \"kubernetes.io/projected/62b08d58-5a82-4492-93e1-eb0df006298c-kube-api-access-62z6m\") on node \"crc\" DevicePath \"\"" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.189326 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.189340 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b08d58-5a82-4492-93e1-eb0df006298c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.587521 4906 generic.go:334] "Generic (PLEG): container finished" podID="62b08d58-5a82-4492-93e1-eb0df006298c" containerID="fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf" exitCode=0 Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.587563 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxk2r" event={"ID":"62b08d58-5a82-4492-93e1-eb0df006298c","Type":"ContainerDied","Data":"fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf"} Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.587592 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxk2r" event={"ID":"62b08d58-5a82-4492-93e1-eb0df006298c","Type":"ContainerDied","Data":"2cf4775ce10678c3d3f047d5efca243d8f468465b642d2db0ea4a1a06b308dc1"} Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.587609 4906 scope.go:117] "RemoveContainer" containerID="fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.587619 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxk2r" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.618859 4906 scope.go:117] "RemoveContainer" containerID="d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.626098 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxk2r"] Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.638262 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxk2r"] Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.647756 4906 scope.go:117] "RemoveContainer" containerID="070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.681106 4906 scope.go:117] "RemoveContainer" containerID="fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf" Nov 23 08:33:07 crc kubenswrapper[4906]: E1123 08:33:07.682034 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf\": container with ID starting with fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf not found: ID does not exist" containerID="fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.682144 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf"} err="failed to get container status \"fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf\": rpc error: code = NotFound desc = could not find container \"fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf\": container with ID starting with fe317846abfa29017d0724faaa53c4f43a1a802123cac21630d6c0f68a38aacf not found: ID does not exist" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.682229 4906 scope.go:117] "RemoveContainer" containerID="d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54" Nov 23 08:33:07 crc kubenswrapper[4906]: E1123 08:33:07.682888 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54\": container with ID starting with d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54 not found: ID does not exist" containerID="d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.682941 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54"} err="failed to get container status \"d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54\": rpc error: code = NotFound desc = could not find container \"d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54\": container with ID starting with d15669d378cab8cfa375b48b6cd50621c51fb695cdc58f92c595b55551ed1a54 not found: ID does not exist" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.682972 4906 scope.go:117] "RemoveContainer" containerID="070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f" Nov 23 08:33:07 crc kubenswrapper[4906]: E1123 08:33:07.683479 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f\": container with ID starting with 070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f not found: ID does not exist" containerID="070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f" Nov 23 08:33:07 crc kubenswrapper[4906]: I1123 08:33:07.683562 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f"} err="failed to get container status \"070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f\": rpc error: code = NotFound desc = could not find container \"070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f\": container with ID starting with 070d489fcdede4065f454a9715d43b6ccd8438c7465e915181f957f24d2ef74f not found: ID does not exist" Nov 23 08:33:09 crc kubenswrapper[4906]: I1123 08:33:09.366776 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" path="/var/lib/kubelet/pods/62b08d58-5a82-4492-93e1-eb0df006298c/volumes" Nov 23 08:33:15 crc kubenswrapper[4906]: I1123 08:33:15.365671 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:33:15 crc kubenswrapper[4906]: E1123 08:33:15.368015 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:33:27 crc kubenswrapper[4906]: I1123 08:33:27.357766 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:33:27 crc kubenswrapper[4906]: E1123 08:33:27.361414 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:33:41 crc kubenswrapper[4906]: I1123 08:33:41.357608 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:33:41 crc kubenswrapper[4906]: E1123 08:33:41.359333 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:33:54 crc kubenswrapper[4906]: I1123 08:33:54.356365 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:33:54 crc kubenswrapper[4906]: E1123 08:33:54.357348 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:34:08 crc kubenswrapper[4906]: I1123 08:34:08.356206 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:34:08 crc kubenswrapper[4906]: E1123 08:34:08.356999 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:34:21 crc kubenswrapper[4906]: I1123 08:34:21.356924 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:34:21 crc kubenswrapper[4906]: E1123 08:34:21.358086 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:34:34 crc kubenswrapper[4906]: I1123 08:34:34.356260 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:34:34 crc kubenswrapper[4906]: E1123 08:34:34.356859 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:34:48 crc kubenswrapper[4906]: I1123 08:34:48.357089 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:34:48 crc kubenswrapper[4906]: E1123 08:34:48.358063 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:35:03 crc kubenswrapper[4906]: I1123 08:35:03.357818 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:35:03 crc kubenswrapper[4906]: I1123 08:35:03.696121 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"aa04aaf31122dd24f5d96cd6a2f7139865ac3380798656efb316b2a8c2dbe038"} Nov 23 08:37:20 crc kubenswrapper[4906]: I1123 08:37:20.945673 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:37:20 crc kubenswrapper[4906]: I1123 08:37:20.946207 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:37:50 crc kubenswrapper[4906]: I1123 08:37:50.945721 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:37:50 crc kubenswrapper[4906]: I1123 08:37:50.946417 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:38:20 crc kubenswrapper[4906]: I1123 08:38:20.945558 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:38:20 crc kubenswrapper[4906]: I1123 08:38:20.946104 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:38:20 crc kubenswrapper[4906]: I1123 08:38:20.946158 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:38:20 crc kubenswrapper[4906]: I1123 08:38:20.946775 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa04aaf31122dd24f5d96cd6a2f7139865ac3380798656efb316b2a8c2dbe038"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:38:20 crc kubenswrapper[4906]: I1123 08:38:20.946845 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://aa04aaf31122dd24f5d96cd6a2f7139865ac3380798656efb316b2a8c2dbe038" gracePeriod=600 Nov 23 08:38:21 crc kubenswrapper[4906]: I1123 08:38:21.533373 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="aa04aaf31122dd24f5d96cd6a2f7139865ac3380798656efb316b2a8c2dbe038" exitCode=0 Nov 23 08:38:21 crc kubenswrapper[4906]: I1123 08:38:21.533450 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"aa04aaf31122dd24f5d96cd6a2f7139865ac3380798656efb316b2a8c2dbe038"} Nov 23 08:38:21 crc kubenswrapper[4906]: I1123 08:38:21.534124 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257"} Nov 23 08:38:21 crc kubenswrapper[4906]: I1123 08:38:21.534154 4906 scope.go:117] "RemoveContainer" containerID="e97e7233a995d5373f3ac1e43c796265b9e2f9d791106837209362f542c55809" Nov 23 08:40:50 crc kubenswrapper[4906]: I1123 08:40:50.945906 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:40:50 crc kubenswrapper[4906]: I1123 08:40:50.946644 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.161624 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cv94d"] Nov 23 08:41:07 crc kubenswrapper[4906]: E1123 08:41:07.163103 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="extract-content" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.163138 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="extract-content" Nov 23 08:41:07 crc kubenswrapper[4906]: E1123 08:41:07.163165 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="extract-utilities" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.163183 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="extract-utilities" Nov 23 08:41:07 crc kubenswrapper[4906]: E1123 08:41:07.163213 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="registry-server" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.163226 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="registry-server" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.163562 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b08d58-5a82-4492-93e1-eb0df006298c" containerName="registry-server" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.166017 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.176664 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cv94d"] Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.305990 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-catalog-content\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.306283 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-utilities\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.306420 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jmmh\" (UniqueName: \"kubernetes.io/projected/c41c0265-9b81-4df0-bcab-d5123053bf22-kube-api-access-8jmmh\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.408352 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-utilities\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.408432 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jmmh\" (UniqueName: \"kubernetes.io/projected/c41c0265-9b81-4df0-bcab-d5123053bf22-kube-api-access-8jmmh\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.408473 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-catalog-content\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.409020 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-catalog-content\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.409020 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-utilities\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.430538 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jmmh\" (UniqueName: \"kubernetes.io/projected/c41c0265-9b81-4df0-bcab-d5123053bf22-kube-api-access-8jmmh\") pod \"redhat-operators-cv94d\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.500024 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:07 crc kubenswrapper[4906]: I1123 08:41:07.920915 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cv94d"] Nov 23 08:41:08 crc kubenswrapper[4906]: I1123 08:41:08.053075 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv94d" event={"ID":"c41c0265-9b81-4df0-bcab-d5123053bf22","Type":"ContainerStarted","Data":"b40aea4b4986e6fd75f0097cd05deb853a5a7a75c096492f525f74b5fc1c1a7a"} Nov 23 08:41:09 crc kubenswrapper[4906]: I1123 08:41:09.061583 4906 generic.go:334] "Generic (PLEG): container finished" podID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerID="c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed" exitCode=0 Nov 23 08:41:09 crc kubenswrapper[4906]: I1123 08:41:09.061619 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv94d" event={"ID":"c41c0265-9b81-4df0-bcab-d5123053bf22","Type":"ContainerDied","Data":"c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed"} Nov 23 08:41:09 crc kubenswrapper[4906]: I1123 08:41:09.063109 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:41:10 crc kubenswrapper[4906]: I1123 08:41:10.076100 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv94d" event={"ID":"c41c0265-9b81-4df0-bcab-d5123053bf22","Type":"ContainerStarted","Data":"bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18"} Nov 23 08:41:11 crc kubenswrapper[4906]: I1123 08:41:11.087437 4906 generic.go:334] "Generic (PLEG): container finished" podID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerID="bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18" exitCode=0 Nov 23 08:41:11 crc kubenswrapper[4906]: I1123 08:41:11.087486 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv94d" event={"ID":"c41c0265-9b81-4df0-bcab-d5123053bf22","Type":"ContainerDied","Data":"bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18"} Nov 23 08:41:12 crc kubenswrapper[4906]: I1123 08:41:12.101833 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv94d" event={"ID":"c41c0265-9b81-4df0-bcab-d5123053bf22","Type":"ContainerStarted","Data":"7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a"} Nov 23 08:41:12 crc kubenswrapper[4906]: I1123 08:41:12.134301 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cv94d" podStartSLOduration=2.620863587 podStartE2EDuration="5.134272865s" podCreationTimestamp="2025-11-23 08:41:07 +0000 UTC" firstStartedPulling="2025-11-23 08:41:09.062912554 +0000 UTC m=+6684.576303857" lastFinishedPulling="2025-11-23 08:41:11.576321822 +0000 UTC m=+6687.089713135" observedRunningTime="2025-11-23 08:41:12.128500221 +0000 UTC m=+6687.641891554" watchObservedRunningTime="2025-11-23 08:41:12.134272865 +0000 UTC m=+6687.647664198" Nov 23 08:41:17 crc kubenswrapper[4906]: I1123 08:41:17.500821 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:17 crc kubenswrapper[4906]: I1123 08:41:17.501434 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:18 crc kubenswrapper[4906]: I1123 08:41:18.561442 4906 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cv94d" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="registry-server" probeResult="failure" output=< Nov 23 08:41:18 crc kubenswrapper[4906]: timeout: failed to connect service ":50051" within 1s Nov 23 08:41:18 crc kubenswrapper[4906]: > Nov 23 08:41:20 crc kubenswrapper[4906]: I1123 08:41:20.945912 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:41:20 crc kubenswrapper[4906]: I1123 08:41:20.946400 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:41:27 crc kubenswrapper[4906]: I1123 08:41:27.557811 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:27 crc kubenswrapper[4906]: I1123 08:41:27.608138 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:27 crc kubenswrapper[4906]: I1123 08:41:27.807456 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cv94d"] Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.271390 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cv94d" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="registry-server" containerID="cri-o://7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a" gracePeriod=2 Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.707173 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.826749 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jmmh\" (UniqueName: \"kubernetes.io/projected/c41c0265-9b81-4df0-bcab-d5123053bf22-kube-api-access-8jmmh\") pod \"c41c0265-9b81-4df0-bcab-d5123053bf22\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.826823 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-catalog-content\") pod \"c41c0265-9b81-4df0-bcab-d5123053bf22\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.826959 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-utilities\") pod \"c41c0265-9b81-4df0-bcab-d5123053bf22\" (UID: \"c41c0265-9b81-4df0-bcab-d5123053bf22\") " Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.828125 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-utilities" (OuterVolumeSpecName: "utilities") pod "c41c0265-9b81-4df0-bcab-d5123053bf22" (UID: "c41c0265-9b81-4df0-bcab-d5123053bf22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.834904 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c41c0265-9b81-4df0-bcab-d5123053bf22-kube-api-access-8jmmh" (OuterVolumeSpecName: "kube-api-access-8jmmh") pod "c41c0265-9b81-4df0-bcab-d5123053bf22" (UID: "c41c0265-9b81-4df0-bcab-d5123053bf22"). InnerVolumeSpecName "kube-api-access-8jmmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.928447 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jmmh\" (UniqueName: \"kubernetes.io/projected/c41c0265-9b81-4df0-bcab-d5123053bf22-kube-api-access-8jmmh\") on node \"crc\" DevicePath \"\"" Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.928486 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:41:29 crc kubenswrapper[4906]: I1123 08:41:29.937771 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c41c0265-9b81-4df0-bcab-d5123053bf22" (UID: "c41c0265-9b81-4df0-bcab-d5123053bf22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.029452 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c41c0265-9b81-4df0-bcab-d5123053bf22-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.282474 4906 generic.go:334] "Generic (PLEG): container finished" podID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerID="7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a" exitCode=0 Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.282534 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv94d" event={"ID":"c41c0265-9b81-4df0-bcab-d5123053bf22","Type":"ContainerDied","Data":"7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a"} Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.282576 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv94d" event={"ID":"c41c0265-9b81-4df0-bcab-d5123053bf22","Type":"ContainerDied","Data":"b40aea4b4986e6fd75f0097cd05deb853a5a7a75c096492f525f74b5fc1c1a7a"} Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.282601 4906 scope.go:117] "RemoveContainer" containerID="7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.282604 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv94d" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.327257 4906 scope.go:117] "RemoveContainer" containerID="bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.329202 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cv94d"] Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.336211 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cv94d"] Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.348757 4906 scope.go:117] "RemoveContainer" containerID="c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.374008 4906 scope.go:117] "RemoveContainer" containerID="7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a" Nov 23 08:41:30 crc kubenswrapper[4906]: E1123 08:41:30.374502 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a\": container with ID starting with 7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a not found: ID does not exist" containerID="7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.374549 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a"} err="failed to get container status \"7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a\": rpc error: code = NotFound desc = could not find container \"7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a\": container with ID starting with 7d90b4d67783e086aff52c75201be072bdfa57df9af6df4ee42f9de04344ac5a not found: ID does not exist" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.374580 4906 scope.go:117] "RemoveContainer" containerID="bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18" Nov 23 08:41:30 crc kubenswrapper[4906]: E1123 08:41:30.374910 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18\": container with ID starting with bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18 not found: ID does not exist" containerID="bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.374934 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18"} err="failed to get container status \"bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18\": rpc error: code = NotFound desc = could not find container \"bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18\": container with ID starting with bea6a5df822bfe4bd9df6e8352070b5f34f282ce57d937c1255c50d1e6399a18 not found: ID does not exist" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.374951 4906 scope.go:117] "RemoveContainer" containerID="c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed" Nov 23 08:41:30 crc kubenswrapper[4906]: E1123 08:41:30.375384 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed\": container with ID starting with c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed not found: ID does not exist" containerID="c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed" Nov 23 08:41:30 crc kubenswrapper[4906]: I1123 08:41:30.375409 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed"} err="failed to get container status \"c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed\": rpc error: code = NotFound desc = could not find container \"c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed\": container with ID starting with c9b1394d532c297c878b01f75e793e4fde74d3266f15aececcec25936cc143ed not found: ID does not exist" Nov 23 08:41:31 crc kubenswrapper[4906]: I1123 08:41:31.374256 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" path="/var/lib/kubelet/pods/c41c0265-9b81-4df0-bcab-d5123053bf22/volumes" Nov 23 08:41:50 crc kubenswrapper[4906]: I1123 08:41:50.945930 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:41:50 crc kubenswrapper[4906]: I1123 08:41:50.946580 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:41:50 crc kubenswrapper[4906]: I1123 08:41:50.946649 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:41:50 crc kubenswrapper[4906]: I1123 08:41:50.947716 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:41:50 crc kubenswrapper[4906]: I1123 08:41:50.947819 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" gracePeriod=600 Nov 23 08:41:51 crc kubenswrapper[4906]: E1123 08:41:51.084952 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:41:51 crc kubenswrapper[4906]: I1123 08:41:51.484502 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" exitCode=0 Nov 23 08:41:51 crc kubenswrapper[4906]: I1123 08:41:51.484569 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257"} Nov 23 08:41:51 crc kubenswrapper[4906]: I1123 08:41:51.484626 4906 scope.go:117] "RemoveContainer" containerID="aa04aaf31122dd24f5d96cd6a2f7139865ac3380798656efb316b2a8c2dbe038" Nov 23 08:41:51 crc kubenswrapper[4906]: I1123 08:41:51.485454 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:41:51 crc kubenswrapper[4906]: E1123 08:41:51.485919 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:42:05 crc kubenswrapper[4906]: I1123 08:42:05.365604 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:42:05 crc kubenswrapper[4906]: E1123 08:42:05.368321 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:42:08 crc kubenswrapper[4906]: I1123 08:42:08.948078 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cczh7"] Nov 23 08:42:08 crc kubenswrapper[4906]: E1123 08:42:08.948991 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="registry-server" Nov 23 08:42:08 crc kubenswrapper[4906]: I1123 08:42:08.949015 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="registry-server" Nov 23 08:42:08 crc kubenswrapper[4906]: E1123 08:42:08.949036 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="extract-utilities" Nov 23 08:42:08 crc kubenswrapper[4906]: I1123 08:42:08.949053 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="extract-utilities" Nov 23 08:42:08 crc kubenswrapper[4906]: E1123 08:42:08.949081 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="extract-content" Nov 23 08:42:08 crc kubenswrapper[4906]: I1123 08:42:08.949095 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="extract-content" Nov 23 08:42:08 crc kubenswrapper[4906]: I1123 08:42:08.949380 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41c0265-9b81-4df0-bcab-d5123053bf22" containerName="registry-server" Nov 23 08:42:08 crc kubenswrapper[4906]: I1123 08:42:08.951295 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:08 crc kubenswrapper[4906]: I1123 08:42:08.972642 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cczh7"] Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.019710 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-catalog-content\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.019786 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x96ww\" (UniqueName: \"kubernetes.io/projected/03a8aa25-500e-4f5a-8918-4dc24db8d476-kube-api-access-x96ww\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.019818 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-utilities\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.120825 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x96ww\" (UniqueName: \"kubernetes.io/projected/03a8aa25-500e-4f5a-8918-4dc24db8d476-kube-api-access-x96ww\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.120920 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-utilities\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.121032 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-catalog-content\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.121654 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-utilities\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.121823 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-catalog-content\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.138603 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x96ww\" (UniqueName: \"kubernetes.io/projected/03a8aa25-500e-4f5a-8918-4dc24db8d476-kube-api-access-x96ww\") pod \"certified-operators-cczh7\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.280124 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:09 crc kubenswrapper[4906]: I1123 08:42:09.768376 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cczh7"] Nov 23 08:42:10 crc kubenswrapper[4906]: I1123 08:42:10.707001 4906 generic.go:334] "Generic (PLEG): container finished" podID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerID="a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f" exitCode=0 Nov 23 08:42:10 crc kubenswrapper[4906]: I1123 08:42:10.707050 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cczh7" event={"ID":"03a8aa25-500e-4f5a-8918-4dc24db8d476","Type":"ContainerDied","Data":"a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f"} Nov 23 08:42:10 crc kubenswrapper[4906]: I1123 08:42:10.707078 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cczh7" event={"ID":"03a8aa25-500e-4f5a-8918-4dc24db8d476","Type":"ContainerStarted","Data":"f69915eb93969eaac7d0572481133172363ae2a91eab437ec9f4ad2b4cf39eef"} Nov 23 08:42:11 crc kubenswrapper[4906]: I1123 08:42:11.718958 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cczh7" event={"ID":"03a8aa25-500e-4f5a-8918-4dc24db8d476","Type":"ContainerStarted","Data":"d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca"} Nov 23 08:42:12 crc kubenswrapper[4906]: I1123 08:42:12.732615 4906 generic.go:334] "Generic (PLEG): container finished" podID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerID="d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca" exitCode=0 Nov 23 08:42:12 crc kubenswrapper[4906]: I1123 08:42:12.732738 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cczh7" event={"ID":"03a8aa25-500e-4f5a-8918-4dc24db8d476","Type":"ContainerDied","Data":"d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca"} Nov 23 08:42:13 crc kubenswrapper[4906]: I1123 08:42:13.742942 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cczh7" event={"ID":"03a8aa25-500e-4f5a-8918-4dc24db8d476","Type":"ContainerStarted","Data":"fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9"} Nov 23 08:42:13 crc kubenswrapper[4906]: I1123 08:42:13.764764 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cczh7" podStartSLOduration=3.27122771 podStartE2EDuration="5.76473815s" podCreationTimestamp="2025-11-23 08:42:08 +0000 UTC" firstStartedPulling="2025-11-23 08:42:10.710850944 +0000 UTC m=+6746.224242237" lastFinishedPulling="2025-11-23 08:42:13.204361364 +0000 UTC m=+6748.717752677" observedRunningTime="2025-11-23 08:42:13.759616034 +0000 UTC m=+6749.273007357" watchObservedRunningTime="2025-11-23 08:42:13.76473815 +0000 UTC m=+6749.278129463" Nov 23 08:42:19 crc kubenswrapper[4906]: I1123 08:42:19.280855 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:19 crc kubenswrapper[4906]: I1123 08:42:19.281939 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:19 crc kubenswrapper[4906]: I1123 08:42:19.338530 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:19 crc kubenswrapper[4906]: I1123 08:42:19.356324 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:42:19 crc kubenswrapper[4906]: E1123 08:42:19.356542 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:42:19 crc kubenswrapper[4906]: I1123 08:42:19.882033 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:19 crc kubenswrapper[4906]: I1123 08:42:19.932473 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cczh7"] Nov 23 08:42:21 crc kubenswrapper[4906]: I1123 08:42:21.825071 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cczh7" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="registry-server" containerID="cri-o://fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9" gracePeriod=2 Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.324970 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.461197 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-utilities\") pod \"03a8aa25-500e-4f5a-8918-4dc24db8d476\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.461318 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x96ww\" (UniqueName: \"kubernetes.io/projected/03a8aa25-500e-4f5a-8918-4dc24db8d476-kube-api-access-x96ww\") pod \"03a8aa25-500e-4f5a-8918-4dc24db8d476\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.461499 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-catalog-content\") pod \"03a8aa25-500e-4f5a-8918-4dc24db8d476\" (UID: \"03a8aa25-500e-4f5a-8918-4dc24db8d476\") " Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.463569 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-utilities" (OuterVolumeSpecName: "utilities") pod "03a8aa25-500e-4f5a-8918-4dc24db8d476" (UID: "03a8aa25-500e-4f5a-8918-4dc24db8d476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.468577 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03a8aa25-500e-4f5a-8918-4dc24db8d476-kube-api-access-x96ww" (OuterVolumeSpecName: "kube-api-access-x96ww") pod "03a8aa25-500e-4f5a-8918-4dc24db8d476" (UID: "03a8aa25-500e-4f5a-8918-4dc24db8d476"). InnerVolumeSpecName "kube-api-access-x96ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.540605 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03a8aa25-500e-4f5a-8918-4dc24db8d476" (UID: "03a8aa25-500e-4f5a-8918-4dc24db8d476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.564428 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.564454 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a8aa25-500e-4f5a-8918-4dc24db8d476-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.564467 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x96ww\" (UniqueName: \"kubernetes.io/projected/03a8aa25-500e-4f5a-8918-4dc24db8d476-kube-api-access-x96ww\") on node \"crc\" DevicePath \"\"" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.832956 4906 generic.go:334] "Generic (PLEG): container finished" podID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerID="fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9" exitCode=0 Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.833002 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cczh7" event={"ID":"03a8aa25-500e-4f5a-8918-4dc24db8d476","Type":"ContainerDied","Data":"fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9"} Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.833030 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cczh7" event={"ID":"03a8aa25-500e-4f5a-8918-4dc24db8d476","Type":"ContainerDied","Data":"f69915eb93969eaac7d0572481133172363ae2a91eab437ec9f4ad2b4cf39eef"} Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.833048 4906 scope.go:117] "RemoveContainer" containerID="fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.833084 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cczh7" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.880666 4906 scope.go:117] "RemoveContainer" containerID="d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.884459 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cczh7"] Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.894040 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cczh7"] Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.904503 4906 scope.go:117] "RemoveContainer" containerID="a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.936637 4906 scope.go:117] "RemoveContainer" containerID="fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9" Nov 23 08:42:22 crc kubenswrapper[4906]: E1123 08:42:22.937226 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9\": container with ID starting with fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9 not found: ID does not exist" containerID="fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.937275 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9"} err="failed to get container status \"fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9\": rpc error: code = NotFound desc = could not find container \"fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9\": container with ID starting with fb360d6bd578771f797d324106b5d29ea064e02f915d832137c9a90822074ba9 not found: ID does not exist" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.937310 4906 scope.go:117] "RemoveContainer" containerID="d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca" Nov 23 08:42:22 crc kubenswrapper[4906]: E1123 08:42:22.937985 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca\": container with ID starting with d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca not found: ID does not exist" containerID="d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.938255 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca"} err="failed to get container status \"d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca\": rpc error: code = NotFound desc = could not find container \"d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca\": container with ID starting with d6b3bdc8b80906628f007b579a8dceea8279c28b84faf6fd175a0bb7c29c0cca not found: ID does not exist" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.938487 4906 scope.go:117] "RemoveContainer" containerID="a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f" Nov 23 08:42:22 crc kubenswrapper[4906]: E1123 08:42:22.939292 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f\": container with ID starting with a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f not found: ID does not exist" containerID="a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f" Nov 23 08:42:22 crc kubenswrapper[4906]: I1123 08:42:22.939341 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f"} err="failed to get container status \"a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f\": rpc error: code = NotFound desc = could not find container \"a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f\": container with ID starting with a0929b67a66bb201838d107a0d75ade7e50cb4637f3bc6af83481dce520c065f not found: ID does not exist" Nov 23 08:42:23 crc kubenswrapper[4906]: I1123 08:42:23.368032 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" path="/var/lib/kubelet/pods/03a8aa25-500e-4f5a-8918-4dc24db8d476/volumes" Nov 23 08:42:30 crc kubenswrapper[4906]: I1123 08:42:30.357167 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:42:30 crc kubenswrapper[4906]: E1123 08:42:30.359835 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:42:45 crc kubenswrapper[4906]: I1123 08:42:45.365335 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:42:45 crc kubenswrapper[4906]: E1123 08:42:45.366421 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:42:58 crc kubenswrapper[4906]: I1123 08:42:58.356475 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:42:58 crc kubenswrapper[4906]: E1123 08:42:58.357093 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:43:10 crc kubenswrapper[4906]: I1123 08:43:10.356394 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:43:10 crc kubenswrapper[4906]: E1123 08:43:10.357411 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:43:21 crc kubenswrapper[4906]: I1123 08:43:21.357525 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:43:21 crc kubenswrapper[4906]: E1123 08:43:21.358955 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:43:35 crc kubenswrapper[4906]: I1123 08:43:35.363745 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:43:35 crc kubenswrapper[4906]: E1123 08:43:35.364785 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:43:46 crc kubenswrapper[4906]: I1123 08:43:46.356668 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:43:46 crc kubenswrapper[4906]: E1123 08:43:46.357415 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:43:58 crc kubenswrapper[4906]: I1123 08:43:58.357930 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:43:58 crc kubenswrapper[4906]: E1123 08:43:58.359251 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:44:12 crc kubenswrapper[4906]: I1123 08:44:12.356124 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:44:12 crc kubenswrapper[4906]: E1123 08:44:12.358207 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:44:24 crc kubenswrapper[4906]: I1123 08:44:24.357011 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:44:24 crc kubenswrapper[4906]: E1123 08:44:24.357657 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:44:35 crc kubenswrapper[4906]: I1123 08:44:35.362721 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:44:35 crc kubenswrapper[4906]: E1123 08:44:35.363891 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:44:49 crc kubenswrapper[4906]: I1123 08:44:49.357042 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:44:49 crc kubenswrapper[4906]: E1123 08:44:49.360254 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.173304 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4"] Nov 23 08:45:00 crc kubenswrapper[4906]: E1123 08:45:00.174085 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="extract-utilities" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.174100 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="extract-utilities" Nov 23 08:45:00 crc kubenswrapper[4906]: E1123 08:45:00.174116 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="extract-content" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.174122 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="extract-content" Nov 23 08:45:00 crc kubenswrapper[4906]: E1123 08:45:00.174144 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="registry-server" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.174152 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="registry-server" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.174296 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a8aa25-500e-4f5a-8918-4dc24db8d476" containerName="registry-server" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.174771 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.177636 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.178072 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.186977 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4"] Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.295559 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da7d9961-ccfd-4296-ba08-a323df54a01d-config-volume\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.295655 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da7d9961-ccfd-4296-ba08-a323df54a01d-secret-volume\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.295685 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js9pl\" (UniqueName: \"kubernetes.io/projected/da7d9961-ccfd-4296-ba08-a323df54a01d-kube-api-access-js9pl\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.397645 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da7d9961-ccfd-4296-ba08-a323df54a01d-secret-volume\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.398188 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js9pl\" (UniqueName: \"kubernetes.io/projected/da7d9961-ccfd-4296-ba08-a323df54a01d-kube-api-access-js9pl\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.398315 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da7d9961-ccfd-4296-ba08-a323df54a01d-config-volume\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.399236 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da7d9961-ccfd-4296-ba08-a323df54a01d-config-volume\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.415898 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da7d9961-ccfd-4296-ba08-a323df54a01d-secret-volume\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.419763 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js9pl\" (UniqueName: \"kubernetes.io/projected/da7d9961-ccfd-4296-ba08-a323df54a01d-kube-api-access-js9pl\") pod \"collect-profiles-29398125-j8kk4\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.491836 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:00 crc kubenswrapper[4906]: I1123 08:45:00.922016 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4"] Nov 23 08:45:01 crc kubenswrapper[4906]: I1123 08:45:01.374561 4906 generic.go:334] "Generic (PLEG): container finished" podID="da7d9961-ccfd-4296-ba08-a323df54a01d" containerID="8c68f46365fb3b41c41a9b0c9f0df2632fde56b1e574c1bd9fd2090109a1c0b2" exitCode=0 Nov 23 08:45:01 crc kubenswrapper[4906]: I1123 08:45:01.374607 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" event={"ID":"da7d9961-ccfd-4296-ba08-a323df54a01d","Type":"ContainerDied","Data":"8c68f46365fb3b41c41a9b0c9f0df2632fde56b1e574c1bd9fd2090109a1c0b2"} Nov 23 08:45:01 crc kubenswrapper[4906]: I1123 08:45:01.374644 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" event={"ID":"da7d9961-ccfd-4296-ba08-a323df54a01d","Type":"ContainerStarted","Data":"5dcfb1d1db8e612b708440212859a598631d01cda9f68fe6fb18fa0d6e36a9d0"} Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.069741 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mhxkh"] Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.073380 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.081319 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhxkh"] Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.225986 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-catalog-content\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.226106 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-utilities\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.226135 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgsgl\" (UniqueName: \"kubernetes.io/projected/531bae01-da69-4a41-9249-447d40aa8b99-kube-api-access-sgsgl\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.327439 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-utilities\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.327491 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgsgl\" (UniqueName: \"kubernetes.io/projected/531bae01-da69-4a41-9249-447d40aa8b99-kube-api-access-sgsgl\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.327517 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-catalog-content\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.328038 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-catalog-content\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.328689 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-utilities\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.348496 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgsgl\" (UniqueName: \"kubernetes.io/projected/531bae01-da69-4a41-9249-447d40aa8b99-kube-api-access-sgsgl\") pod \"community-operators-mhxkh\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.432812 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.722828 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.836785 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da7d9961-ccfd-4296-ba08-a323df54a01d-config-volume\") pod \"da7d9961-ccfd-4296-ba08-a323df54a01d\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.836856 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da7d9961-ccfd-4296-ba08-a323df54a01d-secret-volume\") pod \"da7d9961-ccfd-4296-ba08-a323df54a01d\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.836902 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js9pl\" (UniqueName: \"kubernetes.io/projected/da7d9961-ccfd-4296-ba08-a323df54a01d-kube-api-access-js9pl\") pod \"da7d9961-ccfd-4296-ba08-a323df54a01d\" (UID: \"da7d9961-ccfd-4296-ba08-a323df54a01d\") " Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.838383 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da7d9961-ccfd-4296-ba08-a323df54a01d-config-volume" (OuterVolumeSpecName: "config-volume") pod "da7d9961-ccfd-4296-ba08-a323df54a01d" (UID: "da7d9961-ccfd-4296-ba08-a323df54a01d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.842148 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7d9961-ccfd-4296-ba08-a323df54a01d-kube-api-access-js9pl" (OuterVolumeSpecName: "kube-api-access-js9pl") pod "da7d9961-ccfd-4296-ba08-a323df54a01d" (UID: "da7d9961-ccfd-4296-ba08-a323df54a01d"). InnerVolumeSpecName "kube-api-access-js9pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.842268 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7d9961-ccfd-4296-ba08-a323df54a01d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "da7d9961-ccfd-4296-ba08-a323df54a01d" (UID: "da7d9961-ccfd-4296-ba08-a323df54a01d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.918489 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhxkh"] Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.938006 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js9pl\" (UniqueName: \"kubernetes.io/projected/da7d9961-ccfd-4296-ba08-a323df54a01d-kube-api-access-js9pl\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.938037 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da7d9961-ccfd-4296-ba08-a323df54a01d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:02 crc kubenswrapper[4906]: I1123 08:45:02.938046 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/da7d9961-ccfd-4296-ba08-a323df54a01d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.357789 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:45:03 crc kubenswrapper[4906]: E1123 08:45:03.358079 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.389517 4906 generic.go:334] "Generic (PLEG): container finished" podID="531bae01-da69-4a41-9249-447d40aa8b99" containerID="fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8" exitCode=0 Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.389570 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhxkh" event={"ID":"531bae01-da69-4a41-9249-447d40aa8b99","Type":"ContainerDied","Data":"fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8"} Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.389634 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhxkh" event={"ID":"531bae01-da69-4a41-9249-447d40aa8b99","Type":"ContainerStarted","Data":"ea07249bd5ab3f5478dca1df097cc98211cd2bfee58a553feab0bd44409b319a"} Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.392190 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" event={"ID":"da7d9961-ccfd-4296-ba08-a323df54a01d","Type":"ContainerDied","Data":"5dcfb1d1db8e612b708440212859a598631d01cda9f68fe6fb18fa0d6e36a9d0"} Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.392219 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dcfb1d1db8e612b708440212859a598631d01cda9f68fe6fb18fa0d6e36a9d0" Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.392272 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398125-j8kk4" Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.784628 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7"] Nov 23 08:45:03 crc kubenswrapper[4906]: I1123 08:45:03.788662 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398080-stbg7"] Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.269159 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5zq9f"] Nov 23 08:45:04 crc kubenswrapper[4906]: E1123 08:45:04.269540 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7d9961-ccfd-4296-ba08-a323df54a01d" containerName="collect-profiles" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.269561 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7d9961-ccfd-4296-ba08-a323df54a01d" containerName="collect-profiles" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.269964 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7d9961-ccfd-4296-ba08-a323df54a01d" containerName="collect-profiles" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.275206 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.292060 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zq9f"] Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.358590 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlcgp\" (UniqueName: \"kubernetes.io/projected/b3379960-484f-42e5-af88-90bf3a37150a-kube-api-access-qlcgp\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.358673 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-utilities\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.358716 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-catalog-content\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.400823 4906 generic.go:334] "Generic (PLEG): container finished" podID="531bae01-da69-4a41-9249-447d40aa8b99" containerID="ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e" exitCode=0 Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.400866 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhxkh" event={"ID":"531bae01-da69-4a41-9249-447d40aa8b99","Type":"ContainerDied","Data":"ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e"} Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.460086 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlcgp\" (UniqueName: \"kubernetes.io/projected/b3379960-484f-42e5-af88-90bf3a37150a-kube-api-access-qlcgp\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.460193 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-utilities\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.460236 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-catalog-content\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.460710 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-utilities\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.460748 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-catalog-content\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.477583 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlcgp\" (UniqueName: \"kubernetes.io/projected/b3379960-484f-42e5-af88-90bf3a37150a-kube-api-access-qlcgp\") pod \"redhat-marketplace-5zq9f\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:04 crc kubenswrapper[4906]: I1123 08:45:04.636197 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:05 crc kubenswrapper[4906]: I1123 08:45:05.491988 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d28ea74-f343-4b45-a4a0-56dad68591db" path="/var/lib/kubelet/pods/1d28ea74-f343-4b45-a4a0-56dad68591db/volumes" Nov 23 08:45:05 crc kubenswrapper[4906]: I1123 08:45:05.494657 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zq9f"] Nov 23 08:45:05 crc kubenswrapper[4906]: W1123 08:45:05.522383 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3379960_484f_42e5_af88_90bf3a37150a.slice/crio-7bbad9bc97deb2d52edd3f36dfa8b47a10cea458f1bdd9a86ba56fc677426313 WatchSource:0}: Error finding container 7bbad9bc97deb2d52edd3f36dfa8b47a10cea458f1bdd9a86ba56fc677426313: Status 404 returned error can't find the container with id 7bbad9bc97deb2d52edd3f36dfa8b47a10cea458f1bdd9a86ba56fc677426313 Nov 23 08:45:06 crc kubenswrapper[4906]: I1123 08:45:06.485931 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhxkh" event={"ID":"531bae01-da69-4a41-9249-447d40aa8b99","Type":"ContainerStarted","Data":"c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58"} Nov 23 08:45:06 crc kubenswrapper[4906]: I1123 08:45:06.489058 4906 generic.go:334] "Generic (PLEG): container finished" podID="b3379960-484f-42e5-af88-90bf3a37150a" containerID="19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174" exitCode=0 Nov 23 08:45:06 crc kubenswrapper[4906]: I1123 08:45:06.489110 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zq9f" event={"ID":"b3379960-484f-42e5-af88-90bf3a37150a","Type":"ContainerDied","Data":"19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174"} Nov 23 08:45:06 crc kubenswrapper[4906]: I1123 08:45:06.489142 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zq9f" event={"ID":"b3379960-484f-42e5-af88-90bf3a37150a","Type":"ContainerStarted","Data":"7bbad9bc97deb2d52edd3f36dfa8b47a10cea458f1bdd9a86ba56fc677426313"} Nov 23 08:45:06 crc kubenswrapper[4906]: I1123 08:45:06.505065 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mhxkh" podStartSLOduration=2.397513167 podStartE2EDuration="4.505046612s" podCreationTimestamp="2025-11-23 08:45:02 +0000 UTC" firstStartedPulling="2025-11-23 08:45:03.391704004 +0000 UTC m=+6918.905095307" lastFinishedPulling="2025-11-23 08:45:05.499237449 +0000 UTC m=+6921.012628752" observedRunningTime="2025-11-23 08:45:06.502192786 +0000 UTC m=+6922.015584099" watchObservedRunningTime="2025-11-23 08:45:06.505046612 +0000 UTC m=+6922.018437925" Nov 23 08:45:07 crc kubenswrapper[4906]: I1123 08:45:07.497802 4906 generic.go:334] "Generic (PLEG): container finished" podID="b3379960-484f-42e5-af88-90bf3a37150a" containerID="8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367" exitCode=0 Nov 23 08:45:07 crc kubenswrapper[4906]: I1123 08:45:07.497896 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zq9f" event={"ID":"b3379960-484f-42e5-af88-90bf3a37150a","Type":"ContainerDied","Data":"8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367"} Nov 23 08:45:08 crc kubenswrapper[4906]: I1123 08:45:08.509067 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zq9f" event={"ID":"b3379960-484f-42e5-af88-90bf3a37150a","Type":"ContainerStarted","Data":"24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa"} Nov 23 08:45:08 crc kubenswrapper[4906]: I1123 08:45:08.535001 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5zq9f" podStartSLOduration=3.110992926 podStartE2EDuration="4.534980869s" podCreationTimestamp="2025-11-23 08:45:04 +0000 UTC" firstStartedPulling="2025-11-23 08:45:06.490619347 +0000 UTC m=+6922.004010650" lastFinishedPulling="2025-11-23 08:45:07.91460729 +0000 UTC m=+6923.427998593" observedRunningTime="2025-11-23 08:45:08.531809425 +0000 UTC m=+6924.045200728" watchObservedRunningTime="2025-11-23 08:45:08.534980869 +0000 UTC m=+6924.048372172" Nov 23 08:45:12 crc kubenswrapper[4906]: I1123 08:45:12.434179 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:12 crc kubenswrapper[4906]: I1123 08:45:12.434537 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:12 crc kubenswrapper[4906]: I1123 08:45:12.495190 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:12 crc kubenswrapper[4906]: I1123 08:45:12.579319 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:12 crc kubenswrapper[4906]: I1123 08:45:12.857515 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhxkh"] Nov 23 08:45:14 crc kubenswrapper[4906]: I1123 08:45:14.357138 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:45:14 crc kubenswrapper[4906]: E1123 08:45:14.357377 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:45:14 crc kubenswrapper[4906]: I1123 08:45:14.552713 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mhxkh" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="registry-server" containerID="cri-o://c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58" gracePeriod=2 Nov 23 08:45:14 crc kubenswrapper[4906]: I1123 08:45:14.636621 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:14 crc kubenswrapper[4906]: I1123 08:45:14.636794 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:14 crc kubenswrapper[4906]: I1123 08:45:14.682737 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:14 crc kubenswrapper[4906]: I1123 08:45:14.959240 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.102855 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-catalog-content\") pod \"531bae01-da69-4a41-9249-447d40aa8b99\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.102918 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-utilities\") pod \"531bae01-da69-4a41-9249-447d40aa8b99\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.102997 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgsgl\" (UniqueName: \"kubernetes.io/projected/531bae01-da69-4a41-9249-447d40aa8b99-kube-api-access-sgsgl\") pod \"531bae01-da69-4a41-9249-447d40aa8b99\" (UID: \"531bae01-da69-4a41-9249-447d40aa8b99\") " Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.104190 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-utilities" (OuterVolumeSpecName: "utilities") pod "531bae01-da69-4a41-9249-447d40aa8b99" (UID: "531bae01-da69-4a41-9249-447d40aa8b99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.111307 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/531bae01-da69-4a41-9249-447d40aa8b99-kube-api-access-sgsgl" (OuterVolumeSpecName: "kube-api-access-sgsgl") pod "531bae01-da69-4a41-9249-447d40aa8b99" (UID: "531bae01-da69-4a41-9249-447d40aa8b99"). InnerVolumeSpecName "kube-api-access-sgsgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.155865 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "531bae01-da69-4a41-9249-447d40aa8b99" (UID: "531bae01-da69-4a41-9249-447d40aa8b99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.204021 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgsgl\" (UniqueName: \"kubernetes.io/projected/531bae01-da69-4a41-9249-447d40aa8b99-kube-api-access-sgsgl\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.204060 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.204069 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531bae01-da69-4a41-9249-447d40aa8b99-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.563345 4906 generic.go:334] "Generic (PLEG): container finished" podID="531bae01-da69-4a41-9249-447d40aa8b99" containerID="c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58" exitCode=0 Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.563444 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhxkh" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.563447 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhxkh" event={"ID":"531bae01-da69-4a41-9249-447d40aa8b99","Type":"ContainerDied","Data":"c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58"} Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.563525 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhxkh" event={"ID":"531bae01-da69-4a41-9249-447d40aa8b99","Type":"ContainerDied","Data":"ea07249bd5ab3f5478dca1df097cc98211cd2bfee58a553feab0bd44409b319a"} Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.563563 4906 scope.go:117] "RemoveContainer" containerID="c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.592844 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhxkh"] Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.592896 4906 scope.go:117] "RemoveContainer" containerID="ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.594632 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mhxkh"] Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.609450 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.614517 4906 scope.go:117] "RemoveContainer" containerID="fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.641445 4906 scope.go:117] "RemoveContainer" containerID="c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58" Nov 23 08:45:15 crc kubenswrapper[4906]: E1123 08:45:15.642009 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58\": container with ID starting with c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58 not found: ID does not exist" containerID="c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.642045 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58"} err="failed to get container status \"c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58\": rpc error: code = NotFound desc = could not find container \"c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58\": container with ID starting with c446a7f8a68ec06b9eb4a222ac5272cd269935781177ed0342f0e71fccc09f58 not found: ID does not exist" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.642067 4906 scope.go:117] "RemoveContainer" containerID="ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e" Nov 23 08:45:15 crc kubenswrapper[4906]: E1123 08:45:15.642394 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e\": container with ID starting with ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e not found: ID does not exist" containerID="ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.642417 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e"} err="failed to get container status \"ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e\": rpc error: code = NotFound desc = could not find container \"ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e\": container with ID starting with ff183b72c46e7604b6b13a11ef4f72cc87b3a565f3842069570b56386b651c2e not found: ID does not exist" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.642430 4906 scope.go:117] "RemoveContainer" containerID="fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8" Nov 23 08:45:15 crc kubenswrapper[4906]: E1123 08:45:15.642697 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8\": container with ID starting with fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8 not found: ID does not exist" containerID="fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8" Nov 23 08:45:15 crc kubenswrapper[4906]: I1123 08:45:15.642722 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8"} err="failed to get container status \"fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8\": rpc error: code = NotFound desc = could not find container \"fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8\": container with ID starting with fedc8b5dafe24dc32ea5d004d923cdb1978d7b56eba50c4a382a7ce85b3ad6c8 not found: ID does not exist" Nov 23 08:45:17 crc kubenswrapper[4906]: I1123 08:45:17.368209 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="531bae01-da69-4a41-9249-447d40aa8b99" path="/var/lib/kubelet/pods/531bae01-da69-4a41-9249-447d40aa8b99/volumes" Nov 23 08:45:17 crc kubenswrapper[4906]: I1123 08:45:17.861654 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zq9f"] Nov 23 08:45:18 crc kubenswrapper[4906]: I1123 08:45:18.586057 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5zq9f" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="registry-server" containerID="cri-o://24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa" gracePeriod=2 Nov 23 08:45:18 crc kubenswrapper[4906]: I1123 08:45:18.979479 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.158997 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-utilities\") pod \"b3379960-484f-42e5-af88-90bf3a37150a\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.159061 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlcgp\" (UniqueName: \"kubernetes.io/projected/b3379960-484f-42e5-af88-90bf3a37150a-kube-api-access-qlcgp\") pod \"b3379960-484f-42e5-af88-90bf3a37150a\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.159152 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-catalog-content\") pod \"b3379960-484f-42e5-af88-90bf3a37150a\" (UID: \"b3379960-484f-42e5-af88-90bf3a37150a\") " Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.160057 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-utilities" (OuterVolumeSpecName: "utilities") pod "b3379960-484f-42e5-af88-90bf3a37150a" (UID: "b3379960-484f-42e5-af88-90bf3a37150a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.173978 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3379960-484f-42e5-af88-90bf3a37150a-kube-api-access-qlcgp" (OuterVolumeSpecName: "kube-api-access-qlcgp") pod "b3379960-484f-42e5-af88-90bf3a37150a" (UID: "b3379960-484f-42e5-af88-90bf3a37150a"). InnerVolumeSpecName "kube-api-access-qlcgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.177586 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3379960-484f-42e5-af88-90bf3a37150a" (UID: "b3379960-484f-42e5-af88-90bf3a37150a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.260190 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.260230 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlcgp\" (UniqueName: \"kubernetes.io/projected/b3379960-484f-42e5-af88-90bf3a37150a-kube-api-access-qlcgp\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.260245 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3379960-484f-42e5-af88-90bf3a37150a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.595030 4906 generic.go:334] "Generic (PLEG): container finished" podID="b3379960-484f-42e5-af88-90bf3a37150a" containerID="24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa" exitCode=0 Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.595092 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zq9f" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.595091 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zq9f" event={"ID":"b3379960-484f-42e5-af88-90bf3a37150a","Type":"ContainerDied","Data":"24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa"} Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.595144 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zq9f" event={"ID":"b3379960-484f-42e5-af88-90bf3a37150a","Type":"ContainerDied","Data":"7bbad9bc97deb2d52edd3f36dfa8b47a10cea458f1bdd9a86ba56fc677426313"} Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.595167 4906 scope.go:117] "RemoveContainer" containerID="24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.617925 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zq9f"] Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.618534 4906 scope.go:117] "RemoveContainer" containerID="8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.623282 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zq9f"] Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.634096 4906 scope.go:117] "RemoveContainer" containerID="19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.657946 4906 scope.go:117] "RemoveContainer" containerID="24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa" Nov 23 08:45:19 crc kubenswrapper[4906]: E1123 08:45:19.658454 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa\": container with ID starting with 24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa not found: ID does not exist" containerID="24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.658491 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa"} err="failed to get container status \"24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa\": rpc error: code = NotFound desc = could not find container \"24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa\": container with ID starting with 24d0591701132d340d59e57257eede59467989965e7238c03f8ef53c79b942fa not found: ID does not exist" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.658516 4906 scope.go:117] "RemoveContainer" containerID="8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367" Nov 23 08:45:19 crc kubenswrapper[4906]: E1123 08:45:19.658859 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367\": container with ID starting with 8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367 not found: ID does not exist" containerID="8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.658971 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367"} err="failed to get container status \"8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367\": rpc error: code = NotFound desc = could not find container \"8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367\": container with ID starting with 8081cbf3129aa12e5e1419076cff390c3959561e42ef47f41c091296e93ce367 not found: ID does not exist" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.659061 4906 scope.go:117] "RemoveContainer" containerID="19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174" Nov 23 08:45:19 crc kubenswrapper[4906]: E1123 08:45:19.659620 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174\": container with ID starting with 19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174 not found: ID does not exist" containerID="19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174" Nov 23 08:45:19 crc kubenswrapper[4906]: I1123 08:45:19.659749 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174"} err="failed to get container status \"19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174\": rpc error: code = NotFound desc = could not find container \"19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174\": container with ID starting with 19eb17bedf2a5bec5b1dea91016d32341a691b3b440673408130cbf407b19174 not found: ID does not exist" Nov 23 08:45:21 crc kubenswrapper[4906]: I1123 08:45:21.370414 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3379960-484f-42e5-af88-90bf3a37150a" path="/var/lib/kubelet/pods/b3379960-484f-42e5-af88-90bf3a37150a/volumes" Nov 23 08:45:26 crc kubenswrapper[4906]: I1123 08:45:26.356259 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:45:26 crc kubenswrapper[4906]: E1123 08:45:26.356815 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:45:37 crc kubenswrapper[4906]: I1123 08:45:37.357762 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:45:37 crc kubenswrapper[4906]: E1123 08:45:37.358617 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:45:51 crc kubenswrapper[4906]: I1123 08:45:51.357175 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:45:51 crc kubenswrapper[4906]: E1123 08:45:51.357992 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:45:59 crc kubenswrapper[4906]: I1123 08:45:59.074626 4906 scope.go:117] "RemoveContainer" containerID="3942801e1ce014b645ae0f5047f24ee84d2b3da04d1b40a9792d80de47a708a7" Nov 23 08:46:05 crc kubenswrapper[4906]: I1123 08:46:05.367400 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:46:05 crc kubenswrapper[4906]: E1123 08:46:05.368787 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:46:18 crc kubenswrapper[4906]: I1123 08:46:18.357382 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:46:18 crc kubenswrapper[4906]: E1123 08:46:18.358363 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:46:29 crc kubenswrapper[4906]: I1123 08:46:29.356614 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:46:29 crc kubenswrapper[4906]: E1123 08:46:29.357370 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:46:40 crc kubenswrapper[4906]: I1123 08:46:40.358437 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:46:40 crc kubenswrapper[4906]: E1123 08:46:40.359832 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:46:55 crc kubenswrapper[4906]: I1123 08:46:55.360640 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:46:56 crc kubenswrapper[4906]: I1123 08:46:56.439315 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"a9974406ae09dc3f602193a954c2d90b2a8b06a05979871ce4e1276b10745d47"} Nov 23 08:49:20 crc kubenswrapper[4906]: I1123 08:49:20.945929 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:49:20 crc kubenswrapper[4906]: I1123 08:49:20.946614 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:49:50 crc kubenswrapper[4906]: I1123 08:49:50.946173 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:49:50 crc kubenswrapper[4906]: I1123 08:49:50.947445 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:50:20 crc kubenswrapper[4906]: I1123 08:50:20.945794 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:50:20 crc kubenswrapper[4906]: I1123 08:50:20.946633 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:50:20 crc kubenswrapper[4906]: I1123 08:50:20.946752 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:50:20 crc kubenswrapper[4906]: I1123 08:50:20.947818 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9974406ae09dc3f602193a954c2d90b2a8b06a05979871ce4e1276b10745d47"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:50:20 crc kubenswrapper[4906]: I1123 08:50:20.947940 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://a9974406ae09dc3f602193a954c2d90b2a8b06a05979871ce4e1276b10745d47" gracePeriod=600 Nov 23 08:50:21 crc kubenswrapper[4906]: I1123 08:50:21.295471 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="a9974406ae09dc3f602193a954c2d90b2a8b06a05979871ce4e1276b10745d47" exitCode=0 Nov 23 08:50:21 crc kubenswrapper[4906]: I1123 08:50:21.295556 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"a9974406ae09dc3f602193a954c2d90b2a8b06a05979871ce4e1276b10745d47"} Nov 23 08:50:21 crc kubenswrapper[4906]: I1123 08:50:21.296101 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96"} Nov 23 08:50:21 crc kubenswrapper[4906]: I1123 08:50:21.296132 4906 scope.go:117] "RemoveContainer" containerID="362069712c50348aebb974e323c534a6b09dda1f825670aaa16ac12197efa257" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.921565 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ftrhr"] Nov 23 08:52:12 crc kubenswrapper[4906]: E1123 08:52:12.922330 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="extract-utilities" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922344 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="extract-utilities" Nov 23 08:52:12 crc kubenswrapper[4906]: E1123 08:52:12.922357 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="extract-content" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922362 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="extract-content" Nov 23 08:52:12 crc kubenswrapper[4906]: E1123 08:52:12.922373 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="registry-server" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922379 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="registry-server" Nov 23 08:52:12 crc kubenswrapper[4906]: E1123 08:52:12.922389 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="registry-server" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922395 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="registry-server" Nov 23 08:52:12 crc kubenswrapper[4906]: E1123 08:52:12.922409 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="extract-content" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922414 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="extract-content" Nov 23 08:52:12 crc kubenswrapper[4906]: E1123 08:52:12.922424 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="extract-utilities" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922430 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="extract-utilities" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922564 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="531bae01-da69-4a41-9249-447d40aa8b99" containerName="registry-server" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.922575 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3379960-484f-42e5-af88-90bf3a37150a" containerName="registry-server" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.923612 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:12 crc kubenswrapper[4906]: I1123 08:52:12.939639 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ftrhr"] Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.125557 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-utilities\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.125890 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-catalog-content\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.125967 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjlvg\" (UniqueName: \"kubernetes.io/projected/9689e450-fba0-4fc0-909a-c692ff985e1e-kube-api-access-bjlvg\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.226900 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjlvg\" (UniqueName: \"kubernetes.io/projected/9689e450-fba0-4fc0-909a-c692ff985e1e-kube-api-access-bjlvg\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.227038 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-utilities\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.227084 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-catalog-content\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.227601 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-utilities\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.227640 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-catalog-content\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.255551 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjlvg\" (UniqueName: \"kubernetes.io/projected/9689e450-fba0-4fc0-909a-c692ff985e1e-kube-api-access-bjlvg\") pod \"redhat-operators-ftrhr\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.279349 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:13 crc kubenswrapper[4906]: I1123 08:52:13.729283 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ftrhr"] Nov 23 08:52:14 crc kubenswrapper[4906]: I1123 08:52:14.197454 4906 generic.go:334] "Generic (PLEG): container finished" podID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerID="c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf" exitCode=0 Nov 23 08:52:14 crc kubenswrapper[4906]: I1123 08:52:14.197504 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftrhr" event={"ID":"9689e450-fba0-4fc0-909a-c692ff985e1e","Type":"ContainerDied","Data":"c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf"} Nov 23 08:52:14 crc kubenswrapper[4906]: I1123 08:52:14.197536 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftrhr" event={"ID":"9689e450-fba0-4fc0-909a-c692ff985e1e","Type":"ContainerStarted","Data":"1e7c81271b1f61897dac7c2d79b03cab191cbe2e0a26b61f938aeff659937e0d"} Nov 23 08:52:14 crc kubenswrapper[4906]: I1123 08:52:14.199499 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 08:52:15 crc kubenswrapper[4906]: I1123 08:52:15.207163 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftrhr" event={"ID":"9689e450-fba0-4fc0-909a-c692ff985e1e","Type":"ContainerStarted","Data":"db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3"} Nov 23 08:52:16 crc kubenswrapper[4906]: I1123 08:52:16.214626 4906 generic.go:334] "Generic (PLEG): container finished" podID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerID="db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3" exitCode=0 Nov 23 08:52:16 crc kubenswrapper[4906]: I1123 08:52:16.214718 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftrhr" event={"ID":"9689e450-fba0-4fc0-909a-c692ff985e1e","Type":"ContainerDied","Data":"db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3"} Nov 23 08:52:17 crc kubenswrapper[4906]: I1123 08:52:17.223087 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftrhr" event={"ID":"9689e450-fba0-4fc0-909a-c692ff985e1e","Type":"ContainerStarted","Data":"4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf"} Nov 23 08:52:17 crc kubenswrapper[4906]: I1123 08:52:17.240663 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ftrhr" podStartSLOduration=2.826379996 podStartE2EDuration="5.240645638s" podCreationTimestamp="2025-11-23 08:52:12 +0000 UTC" firstStartedPulling="2025-11-23 08:52:14.199186995 +0000 UTC m=+7349.712578298" lastFinishedPulling="2025-11-23 08:52:16.613452627 +0000 UTC m=+7352.126843940" observedRunningTime="2025-11-23 08:52:17.237956156 +0000 UTC m=+7352.751347459" watchObservedRunningTime="2025-11-23 08:52:17.240645638 +0000 UTC m=+7352.754036941" Nov 23 08:52:23 crc kubenswrapper[4906]: I1123 08:52:23.279834 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:23 crc kubenswrapper[4906]: I1123 08:52:23.280484 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:23 crc kubenswrapper[4906]: I1123 08:52:23.327220 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:24 crc kubenswrapper[4906]: I1123 08:52:24.326379 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:24 crc kubenswrapper[4906]: I1123 08:52:24.389513 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ftrhr"] Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.298438 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ftrhr" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="registry-server" containerID="cri-o://4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf" gracePeriod=2 Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.752421 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.864286 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjlvg\" (UniqueName: \"kubernetes.io/projected/9689e450-fba0-4fc0-909a-c692ff985e1e-kube-api-access-bjlvg\") pod \"9689e450-fba0-4fc0-909a-c692ff985e1e\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.864629 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-utilities\") pod \"9689e450-fba0-4fc0-909a-c692ff985e1e\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.864889 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-catalog-content\") pod \"9689e450-fba0-4fc0-909a-c692ff985e1e\" (UID: \"9689e450-fba0-4fc0-909a-c692ff985e1e\") " Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.866106 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-utilities" (OuterVolumeSpecName: "utilities") pod "9689e450-fba0-4fc0-909a-c692ff985e1e" (UID: "9689e450-fba0-4fc0-909a-c692ff985e1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.871254 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9689e450-fba0-4fc0-909a-c692ff985e1e-kube-api-access-bjlvg" (OuterVolumeSpecName: "kube-api-access-bjlvg") pod "9689e450-fba0-4fc0-909a-c692ff985e1e" (UID: "9689e450-fba0-4fc0-909a-c692ff985e1e"). InnerVolumeSpecName "kube-api-access-bjlvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.966135 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjlvg\" (UniqueName: \"kubernetes.io/projected/9689e450-fba0-4fc0-909a-c692ff985e1e-kube-api-access-bjlvg\") on node \"crc\" DevicePath \"\"" Nov 23 08:52:26 crc kubenswrapper[4906]: I1123 08:52:26.966198 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.002509 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9689e450-fba0-4fc0-909a-c692ff985e1e" (UID: "9689e450-fba0-4fc0-909a-c692ff985e1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.068429 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9689e450-fba0-4fc0-909a-c692ff985e1e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.313286 4906 generic.go:334] "Generic (PLEG): container finished" podID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerID="4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf" exitCode=0 Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.313347 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftrhr" event={"ID":"9689e450-fba0-4fc0-909a-c692ff985e1e","Type":"ContainerDied","Data":"4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf"} Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.313387 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftrhr" event={"ID":"9689e450-fba0-4fc0-909a-c692ff985e1e","Type":"ContainerDied","Data":"1e7c81271b1f61897dac7c2d79b03cab191cbe2e0a26b61f938aeff659937e0d"} Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.313407 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftrhr" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.313417 4906 scope.go:117] "RemoveContainer" containerID="4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.352058 4906 scope.go:117] "RemoveContainer" containerID="db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.384273 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ftrhr"] Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.391497 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ftrhr"] Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.398355 4906 scope.go:117] "RemoveContainer" containerID="c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.430397 4906 scope.go:117] "RemoveContainer" containerID="4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf" Nov 23 08:52:27 crc kubenswrapper[4906]: E1123 08:52:27.431096 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf\": container with ID starting with 4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf not found: ID does not exist" containerID="4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.431289 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf"} err="failed to get container status \"4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf\": rpc error: code = NotFound desc = could not find container \"4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf\": container with ID starting with 4c10992ae8558e02c797e80177a893a8658f9daa7d5de306b140e711c9d183bf not found: ID does not exist" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.431370 4906 scope.go:117] "RemoveContainer" containerID="db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3" Nov 23 08:52:27 crc kubenswrapper[4906]: E1123 08:52:27.431883 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3\": container with ID starting with db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3 not found: ID does not exist" containerID="db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.432001 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3"} err="failed to get container status \"db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3\": rpc error: code = NotFound desc = could not find container \"db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3\": container with ID starting with db7242183ebebec9d93ee11f071441a75974eddca9cf9106630322e507d148f3 not found: ID does not exist" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.432050 4906 scope.go:117] "RemoveContainer" containerID="c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf" Nov 23 08:52:27 crc kubenswrapper[4906]: E1123 08:52:27.432571 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf\": container with ID starting with c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf not found: ID does not exist" containerID="c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf" Nov 23 08:52:27 crc kubenswrapper[4906]: I1123 08:52:27.432652 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf"} err="failed to get container status \"c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf\": rpc error: code = NotFound desc = could not find container \"c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf\": container with ID starting with c5f6e260bc5c8c34937c5341f26194b81ada547c9e6cb5966cbfed661999e5bf not found: ID does not exist" Nov 23 08:52:29 crc kubenswrapper[4906]: I1123 08:52:29.372389 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" path="/var/lib/kubelet/pods/9689e450-fba0-4fc0-909a-c692ff985e1e/volumes" Nov 23 08:52:50 crc kubenswrapper[4906]: I1123 08:52:50.945577 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:52:50 crc kubenswrapper[4906]: I1123 08:52:50.946836 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.002981 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jczbw"] Nov 23 08:53:02 crc kubenswrapper[4906]: E1123 08:53:02.004508 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="extract-content" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.004540 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="extract-content" Nov 23 08:53:02 crc kubenswrapper[4906]: E1123 08:53:02.004579 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="extract-utilities" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.004597 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="extract-utilities" Nov 23 08:53:02 crc kubenswrapper[4906]: E1123 08:53:02.004725 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="registry-server" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.004749 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="registry-server" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.006540 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="9689e450-fba0-4fc0-909a-c692ff985e1e" containerName="registry-server" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.010087 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.021383 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jczbw"] Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.059084 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-catalog-content\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.059185 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-utilities\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.059256 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4bhr\" (UniqueName: \"kubernetes.io/projected/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-kube-api-access-k4bhr\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.160124 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-catalog-content\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.160219 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-utilities\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.160274 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4bhr\" (UniqueName: \"kubernetes.io/projected/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-kube-api-access-k4bhr\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.161529 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-utilities\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.161559 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-catalog-content\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.187660 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4bhr\" (UniqueName: \"kubernetes.io/projected/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-kube-api-access-k4bhr\") pod \"certified-operators-jczbw\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.402889 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:02 crc kubenswrapper[4906]: I1123 08:53:02.698380 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jczbw"] Nov 23 08:53:03 crc kubenswrapper[4906]: I1123 08:53:03.655117 4906 generic.go:334] "Generic (PLEG): container finished" podID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerID="22360cd61b2308f2c48e3bd6f308aa87f530371f84990c45b4753e2681440277" exitCode=0 Nov 23 08:53:03 crc kubenswrapper[4906]: I1123 08:53:03.655294 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczbw" event={"ID":"112b2e9b-827b-48ec-bd84-e2b3630b3ae4","Type":"ContainerDied","Data":"22360cd61b2308f2c48e3bd6f308aa87f530371f84990c45b4753e2681440277"} Nov 23 08:53:03 crc kubenswrapper[4906]: I1123 08:53:03.655439 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczbw" event={"ID":"112b2e9b-827b-48ec-bd84-e2b3630b3ae4","Type":"ContainerStarted","Data":"0a81c1b3f7355108074e292ef471e19048355bc4330ff1f8f91f4cd1efaad66b"} Nov 23 08:53:04 crc kubenswrapper[4906]: I1123 08:53:04.665147 4906 generic.go:334] "Generic (PLEG): container finished" podID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerID="2dbd3cd8c28cb02f671ea724130b2e082ba5e2949c0b67077d0f91d3f5f81950" exitCode=0 Nov 23 08:53:04 crc kubenswrapper[4906]: I1123 08:53:04.665371 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczbw" event={"ID":"112b2e9b-827b-48ec-bd84-e2b3630b3ae4","Type":"ContainerDied","Data":"2dbd3cd8c28cb02f671ea724130b2e082ba5e2949c0b67077d0f91d3f5f81950"} Nov 23 08:53:05 crc kubenswrapper[4906]: I1123 08:53:05.674880 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczbw" event={"ID":"112b2e9b-827b-48ec-bd84-e2b3630b3ae4","Type":"ContainerStarted","Data":"d579552fd7f4caf60428c0efb4b5ac01afce807c0cc3ef9e854453c20888da79"} Nov 23 08:53:05 crc kubenswrapper[4906]: I1123 08:53:05.693829 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jczbw" podStartSLOduration=3.270518914 podStartE2EDuration="4.693812288s" podCreationTimestamp="2025-11-23 08:53:01 +0000 UTC" firstStartedPulling="2025-11-23 08:53:03.657623005 +0000 UTC m=+7399.171014348" lastFinishedPulling="2025-11-23 08:53:05.080916379 +0000 UTC m=+7400.594307722" observedRunningTime="2025-11-23 08:53:05.691875057 +0000 UTC m=+7401.205266370" watchObservedRunningTime="2025-11-23 08:53:05.693812288 +0000 UTC m=+7401.207203591" Nov 23 08:53:12 crc kubenswrapper[4906]: I1123 08:53:12.403482 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:12 crc kubenswrapper[4906]: I1123 08:53:12.404429 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:12 crc kubenswrapper[4906]: I1123 08:53:12.496896 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:12 crc kubenswrapper[4906]: I1123 08:53:12.796564 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:12 crc kubenswrapper[4906]: I1123 08:53:12.864113 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jczbw"] Nov 23 08:53:14 crc kubenswrapper[4906]: I1123 08:53:14.752319 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jczbw" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="registry-server" containerID="cri-o://d579552fd7f4caf60428c0efb4b5ac01afce807c0cc3ef9e854453c20888da79" gracePeriod=2 Nov 23 08:53:15 crc kubenswrapper[4906]: I1123 08:53:15.767211 4906 generic.go:334] "Generic (PLEG): container finished" podID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerID="d579552fd7f4caf60428c0efb4b5ac01afce807c0cc3ef9e854453c20888da79" exitCode=0 Nov 23 08:53:15 crc kubenswrapper[4906]: I1123 08:53:15.767313 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczbw" event={"ID":"112b2e9b-827b-48ec-bd84-e2b3630b3ae4","Type":"ContainerDied","Data":"d579552fd7f4caf60428c0efb4b5ac01afce807c0cc3ef9e854453c20888da79"} Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.318790 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.418487 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4bhr\" (UniqueName: \"kubernetes.io/projected/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-kube-api-access-k4bhr\") pod \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.418569 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-utilities\") pod \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.418650 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-catalog-content\") pod \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\" (UID: \"112b2e9b-827b-48ec-bd84-e2b3630b3ae4\") " Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.419586 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-utilities" (OuterVolumeSpecName: "utilities") pod "112b2e9b-827b-48ec-bd84-e2b3630b3ae4" (UID: "112b2e9b-827b-48ec-bd84-e2b3630b3ae4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.443707 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-kube-api-access-k4bhr" (OuterVolumeSpecName: "kube-api-access-k4bhr") pod "112b2e9b-827b-48ec-bd84-e2b3630b3ae4" (UID: "112b2e9b-827b-48ec-bd84-e2b3630b3ae4"). InnerVolumeSpecName "kube-api-access-k4bhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.489569 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "112b2e9b-827b-48ec-bd84-e2b3630b3ae4" (UID: "112b2e9b-827b-48ec-bd84-e2b3630b3ae4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.519673 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.519725 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.519736 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4bhr\" (UniqueName: \"kubernetes.io/projected/112b2e9b-827b-48ec-bd84-e2b3630b3ae4-kube-api-access-k4bhr\") on node \"crc\" DevicePath \"\"" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.776384 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczbw" event={"ID":"112b2e9b-827b-48ec-bd84-e2b3630b3ae4","Type":"ContainerDied","Data":"0a81c1b3f7355108074e292ef471e19048355bc4330ff1f8f91f4cd1efaad66b"} Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.776461 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczbw" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.776765 4906 scope.go:117] "RemoveContainer" containerID="d579552fd7f4caf60428c0efb4b5ac01afce807c0cc3ef9e854453c20888da79" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.795705 4906 scope.go:117] "RemoveContainer" containerID="2dbd3cd8c28cb02f671ea724130b2e082ba5e2949c0b67077d0f91d3f5f81950" Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.817336 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jczbw"] Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.820395 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jczbw"] Nov 23 08:53:16 crc kubenswrapper[4906]: I1123 08:53:16.835003 4906 scope.go:117] "RemoveContainer" containerID="22360cd61b2308f2c48e3bd6f308aa87f530371f84990c45b4753e2681440277" Nov 23 08:53:17 crc kubenswrapper[4906]: I1123 08:53:17.371079 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" path="/var/lib/kubelet/pods/112b2e9b-827b-48ec-bd84-e2b3630b3ae4/volumes" Nov 23 08:53:20 crc kubenswrapper[4906]: I1123 08:53:20.945575 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:53:20 crc kubenswrapper[4906]: I1123 08:53:20.946117 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:53:50 crc kubenswrapper[4906]: I1123 08:53:50.946153 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 08:53:50 crc kubenswrapper[4906]: I1123 08:53:50.946904 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 08:53:50 crc kubenswrapper[4906]: I1123 08:53:50.946973 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 08:53:50 crc kubenswrapper[4906]: I1123 08:53:50.948045 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 08:53:50 crc kubenswrapper[4906]: I1123 08:53:50.948180 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" gracePeriod=600 Nov 23 08:53:51 crc kubenswrapper[4906]: I1123 08:53:51.102989 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" exitCode=0 Nov 23 08:53:51 crc kubenswrapper[4906]: I1123 08:53:51.103071 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96"} Nov 23 08:53:51 crc kubenswrapper[4906]: I1123 08:53:51.103130 4906 scope.go:117] "RemoveContainer" containerID="a9974406ae09dc3f602193a954c2d90b2a8b06a05979871ce4e1276b10745d47" Nov 23 08:53:51 crc kubenswrapper[4906]: E1123 08:53:51.111952 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:53:52 crc kubenswrapper[4906]: I1123 08:53:52.115404 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:53:52 crc kubenswrapper[4906]: E1123 08:53:52.115992 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:54:03 crc kubenswrapper[4906]: I1123 08:54:03.357131 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:54:03 crc kubenswrapper[4906]: E1123 08:54:03.358295 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:54:14 crc kubenswrapper[4906]: I1123 08:54:14.356450 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:54:14 crc kubenswrapper[4906]: E1123 08:54:14.357229 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:54:29 crc kubenswrapper[4906]: I1123 08:54:29.357103 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:54:29 crc kubenswrapper[4906]: E1123 08:54:29.358110 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:54:41 crc kubenswrapper[4906]: I1123 08:54:41.357536 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:54:41 crc kubenswrapper[4906]: E1123 08:54:41.358556 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:54:53 crc kubenswrapper[4906]: I1123 08:54:53.357139 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:54:53 crc kubenswrapper[4906]: E1123 08:54:53.357966 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:55:04 crc kubenswrapper[4906]: I1123 08:55:04.356730 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:55:04 crc kubenswrapper[4906]: E1123 08:55:04.357540 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:55:18 crc kubenswrapper[4906]: I1123 08:55:18.356991 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:55:18 crc kubenswrapper[4906]: E1123 08:55:18.357911 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:55:30 crc kubenswrapper[4906]: I1123 08:55:30.364268 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:55:30 crc kubenswrapper[4906]: E1123 08:55:30.365359 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:55:45 crc kubenswrapper[4906]: I1123 08:55:45.364955 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:55:45 crc kubenswrapper[4906]: E1123 08:55:45.365525 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:55:57 crc kubenswrapper[4906]: I1123 08:55:57.903857 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-84f49"] Nov 23 08:55:57 crc kubenswrapper[4906]: E1123 08:55:57.904802 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="extract-content" Nov 23 08:55:57 crc kubenswrapper[4906]: I1123 08:55:57.904820 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="extract-content" Nov 23 08:55:57 crc kubenswrapper[4906]: E1123 08:55:57.904842 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="registry-server" Nov 23 08:55:57 crc kubenswrapper[4906]: I1123 08:55:57.904850 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="registry-server" Nov 23 08:55:57 crc kubenswrapper[4906]: E1123 08:55:57.904860 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="extract-utilities" Nov 23 08:55:57 crc kubenswrapper[4906]: I1123 08:55:57.904868 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="extract-utilities" Nov 23 08:55:57 crc kubenswrapper[4906]: I1123 08:55:57.905061 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="112b2e9b-827b-48ec-bd84-e2b3630b3ae4" containerName="registry-server" Nov 23 08:55:57 crc kubenswrapper[4906]: I1123 08:55:57.907029 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:57 crc kubenswrapper[4906]: I1123 08:55:57.966656 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84f49"] Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.009373 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-catalog-content\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.009733 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw665\" (UniqueName: \"kubernetes.io/projected/913b889c-b845-47e5-bcee-b0313b95c4b7-kube-api-access-vw665\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.009876 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-utilities\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.111456 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-catalog-content\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.111511 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw665\" (UniqueName: \"kubernetes.io/projected/913b889c-b845-47e5-bcee-b0313b95c4b7-kube-api-access-vw665\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.111559 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-utilities\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.112040 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-catalog-content\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.112060 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-utilities\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.130784 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw665\" (UniqueName: \"kubernetes.io/projected/913b889c-b845-47e5-bcee-b0313b95c4b7-kube-api-access-vw665\") pod \"community-operators-84f49\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.264236 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:55:58 crc kubenswrapper[4906]: I1123 08:55:58.754644 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84f49"] Nov 23 08:55:59 crc kubenswrapper[4906]: I1123 08:55:59.318048 4906 generic.go:334] "Generic (PLEG): container finished" podID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerID="a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e" exitCode=0 Nov 23 08:55:59 crc kubenswrapper[4906]: I1123 08:55:59.318111 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84f49" event={"ID":"913b889c-b845-47e5-bcee-b0313b95c4b7","Type":"ContainerDied","Data":"a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e"} Nov 23 08:55:59 crc kubenswrapper[4906]: I1123 08:55:59.318150 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84f49" event={"ID":"913b889c-b845-47e5-bcee-b0313b95c4b7","Type":"ContainerStarted","Data":"892baffe8174e9d2867fe92247604f2a421065bfb1b83ab8380d27de3739e300"} Nov 23 08:55:59 crc kubenswrapper[4906]: I1123 08:55:59.358086 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:55:59 crc kubenswrapper[4906]: E1123 08:55:59.358544 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:56:00 crc kubenswrapper[4906]: I1123 08:56:00.329140 4906 generic.go:334] "Generic (PLEG): container finished" podID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerID="bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28" exitCode=0 Nov 23 08:56:00 crc kubenswrapper[4906]: I1123 08:56:00.329268 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84f49" event={"ID":"913b889c-b845-47e5-bcee-b0313b95c4b7","Type":"ContainerDied","Data":"bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28"} Nov 23 08:56:01 crc kubenswrapper[4906]: I1123 08:56:01.341145 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84f49" event={"ID":"913b889c-b845-47e5-bcee-b0313b95c4b7","Type":"ContainerStarted","Data":"8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791"} Nov 23 08:56:01 crc kubenswrapper[4906]: I1123 08:56:01.366915 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-84f49" podStartSLOduration=2.977883674 podStartE2EDuration="4.366897692s" podCreationTimestamp="2025-11-23 08:55:57 +0000 UTC" firstStartedPulling="2025-11-23 08:55:59.319796181 +0000 UTC m=+7574.833187524" lastFinishedPulling="2025-11-23 08:56:00.708810239 +0000 UTC m=+7576.222201542" observedRunningTime="2025-11-23 08:56:01.364995922 +0000 UTC m=+7576.878387305" watchObservedRunningTime="2025-11-23 08:56:01.366897692 +0000 UTC m=+7576.880289015" Nov 23 08:56:08 crc kubenswrapper[4906]: I1123 08:56:08.264712 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:56:08 crc kubenswrapper[4906]: I1123 08:56:08.265312 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:56:08 crc kubenswrapper[4906]: I1123 08:56:08.326306 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:56:08 crc kubenswrapper[4906]: I1123 08:56:08.463017 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:56:08 crc kubenswrapper[4906]: I1123 08:56:08.555334 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84f49"] Nov 23 08:56:10 crc kubenswrapper[4906]: I1123 08:56:10.402002 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-84f49" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="registry-server" containerID="cri-o://8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791" gracePeriod=2 Nov 23 08:56:10 crc kubenswrapper[4906]: I1123 08:56:10.798086 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:56:10 crc kubenswrapper[4906]: I1123 08:56:10.947425 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-catalog-content\") pod \"913b889c-b845-47e5-bcee-b0313b95c4b7\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " Nov 23 08:56:10 crc kubenswrapper[4906]: I1123 08:56:10.947529 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw665\" (UniqueName: \"kubernetes.io/projected/913b889c-b845-47e5-bcee-b0313b95c4b7-kube-api-access-vw665\") pod \"913b889c-b845-47e5-bcee-b0313b95c4b7\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " Nov 23 08:56:10 crc kubenswrapper[4906]: I1123 08:56:10.947553 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-utilities\") pod \"913b889c-b845-47e5-bcee-b0313b95c4b7\" (UID: \"913b889c-b845-47e5-bcee-b0313b95c4b7\") " Nov 23 08:56:10 crc kubenswrapper[4906]: I1123 08:56:10.948417 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-utilities" (OuterVolumeSpecName: "utilities") pod "913b889c-b845-47e5-bcee-b0313b95c4b7" (UID: "913b889c-b845-47e5-bcee-b0313b95c4b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:56:10 crc kubenswrapper[4906]: I1123 08:56:10.953224 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913b889c-b845-47e5-bcee-b0313b95c4b7-kube-api-access-vw665" (OuterVolumeSpecName: "kube-api-access-vw665") pod "913b889c-b845-47e5-bcee-b0313b95c4b7" (UID: "913b889c-b845-47e5-bcee-b0313b95c4b7"). InnerVolumeSpecName "kube-api-access-vw665". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.049180 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw665\" (UniqueName: \"kubernetes.io/projected/913b889c-b845-47e5-bcee-b0313b95c4b7-kube-api-access-vw665\") on node \"crc\" DevicePath \"\"" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.049214 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.357553 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:56:11 crc kubenswrapper[4906]: E1123 08:56:11.358101 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.383924 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "913b889c-b845-47e5-bcee-b0313b95c4b7" (UID: "913b889c-b845-47e5-bcee-b0313b95c4b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.413316 4906 generic.go:334] "Generic (PLEG): container finished" podID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerID="8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791" exitCode=0 Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.413374 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84f49" event={"ID":"913b889c-b845-47e5-bcee-b0313b95c4b7","Type":"ContainerDied","Data":"8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791"} Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.413385 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84f49" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.413410 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84f49" event={"ID":"913b889c-b845-47e5-bcee-b0313b95c4b7","Type":"ContainerDied","Data":"892baffe8174e9d2867fe92247604f2a421065bfb1b83ab8380d27de3739e300"} Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.413436 4906 scope.go:117] "RemoveContainer" containerID="8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.456199 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/913b889c-b845-47e5-bcee-b0313b95c4b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.458386 4906 scope.go:117] "RemoveContainer" containerID="bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.459458 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84f49"] Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.468285 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-84f49"] Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.480822 4906 scope.go:117] "RemoveContainer" containerID="a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.506490 4906 scope.go:117] "RemoveContainer" containerID="8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791" Nov 23 08:56:11 crc kubenswrapper[4906]: E1123 08:56:11.507024 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791\": container with ID starting with 8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791 not found: ID does not exist" containerID="8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.507078 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791"} err="failed to get container status \"8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791\": rpc error: code = NotFound desc = could not find container \"8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791\": container with ID starting with 8f41a8e2a0783fa494fa28bb65b11779c3a9799ed16b43d425055f7949654791 not found: ID does not exist" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.507111 4906 scope.go:117] "RemoveContainer" containerID="bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28" Nov 23 08:56:11 crc kubenswrapper[4906]: E1123 08:56:11.507409 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28\": container with ID starting with bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28 not found: ID does not exist" containerID="bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.507456 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28"} err="failed to get container status \"bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28\": rpc error: code = NotFound desc = could not find container \"bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28\": container with ID starting with bec0c5f2479aab1603137e3a2bddc78a1c14f4309f85f7fce4ea1bacf2e0aa28 not found: ID does not exist" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.507484 4906 scope.go:117] "RemoveContainer" containerID="a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e" Nov 23 08:56:11 crc kubenswrapper[4906]: E1123 08:56:11.507954 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e\": container with ID starting with a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e not found: ID does not exist" containerID="a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e" Nov 23 08:56:11 crc kubenswrapper[4906]: I1123 08:56:11.507980 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e"} err="failed to get container status \"a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e\": rpc error: code = NotFound desc = could not find container \"a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e\": container with ID starting with a2b92851dafc7037e040ea15cadfab4f8512955183399ea1fd658350dabb2a8e not found: ID does not exist" Nov 23 08:56:13 crc kubenswrapper[4906]: I1123 08:56:13.364820 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" path="/var/lib/kubelet/pods/913b889c-b845-47e5-bcee-b0313b95c4b7/volumes" Nov 23 08:56:22 crc kubenswrapper[4906]: I1123 08:56:22.356743 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:56:22 crc kubenswrapper[4906]: E1123 08:56:22.357531 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:56:33 crc kubenswrapper[4906]: I1123 08:56:33.356518 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:56:33 crc kubenswrapper[4906]: E1123 08:56:33.357552 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:56:33 crc kubenswrapper[4906]: I1123 08:56:33.987672 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tgk8q"] Nov 23 08:56:33 crc kubenswrapper[4906]: E1123 08:56:33.987970 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="extract-utilities" Nov 23 08:56:33 crc kubenswrapper[4906]: I1123 08:56:33.987985 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="extract-utilities" Nov 23 08:56:33 crc kubenswrapper[4906]: E1123 08:56:33.988007 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="extract-content" Nov 23 08:56:33 crc kubenswrapper[4906]: I1123 08:56:33.988012 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="extract-content" Nov 23 08:56:33 crc kubenswrapper[4906]: E1123 08:56:33.988034 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="registry-server" Nov 23 08:56:33 crc kubenswrapper[4906]: I1123 08:56:33.988040 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="registry-server" Nov 23 08:56:33 crc kubenswrapper[4906]: I1123 08:56:33.988155 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="913b889c-b845-47e5-bcee-b0313b95c4b7" containerName="registry-server" Nov 23 08:56:33 crc kubenswrapper[4906]: I1123 08:56:33.990469 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.015802 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgk8q"] Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.110347 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-catalog-content\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.110697 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-utilities\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.110862 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc4ml\" (UniqueName: \"kubernetes.io/projected/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-kube-api-access-jc4ml\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.211794 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc4ml\" (UniqueName: \"kubernetes.io/projected/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-kube-api-access-jc4ml\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.211913 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-catalog-content\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.211952 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-utilities\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.212612 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-utilities\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.212611 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-catalog-content\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.238484 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc4ml\" (UniqueName: \"kubernetes.io/projected/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-kube-api-access-jc4ml\") pod \"redhat-marketplace-tgk8q\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.309735 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:34 crc kubenswrapper[4906]: I1123 08:56:34.762524 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgk8q"] Nov 23 08:56:35 crc kubenswrapper[4906]: I1123 08:56:35.632096 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgk8q" event={"ID":"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a","Type":"ContainerDied","Data":"dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6"} Nov 23 08:56:35 crc kubenswrapper[4906]: I1123 08:56:35.631913 4906 generic.go:334] "Generic (PLEG): container finished" podID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerID="dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6" exitCode=0 Nov 23 08:56:35 crc kubenswrapper[4906]: I1123 08:56:35.632917 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgk8q" event={"ID":"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a","Type":"ContainerStarted","Data":"2be151dfe3bcfa254c5172c93110c70715b3f3e7fd1fc2210cf4e9c58f874f86"} Nov 23 08:56:37 crc kubenswrapper[4906]: I1123 08:56:37.649985 4906 generic.go:334] "Generic (PLEG): container finished" podID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerID="247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4" exitCode=0 Nov 23 08:56:37 crc kubenswrapper[4906]: I1123 08:56:37.650036 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgk8q" event={"ID":"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a","Type":"ContainerDied","Data":"247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4"} Nov 23 08:56:38 crc kubenswrapper[4906]: I1123 08:56:38.666277 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgk8q" event={"ID":"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a","Type":"ContainerStarted","Data":"7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c"} Nov 23 08:56:38 crc kubenswrapper[4906]: I1123 08:56:38.703713 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tgk8q" podStartSLOduration=3.243772466 podStartE2EDuration="5.703676008s" podCreationTimestamp="2025-11-23 08:56:33 +0000 UTC" firstStartedPulling="2025-11-23 08:56:35.648149498 +0000 UTC m=+7611.161540811" lastFinishedPulling="2025-11-23 08:56:38.10805305 +0000 UTC m=+7613.621444353" observedRunningTime="2025-11-23 08:56:38.698874259 +0000 UTC m=+7614.212265572" watchObservedRunningTime="2025-11-23 08:56:38.703676008 +0000 UTC m=+7614.217067311" Nov 23 08:56:44 crc kubenswrapper[4906]: I1123 08:56:44.311520 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:44 crc kubenswrapper[4906]: I1123 08:56:44.311772 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:44 crc kubenswrapper[4906]: I1123 08:56:44.367241 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:44 crc kubenswrapper[4906]: I1123 08:56:44.769725 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:44 crc kubenswrapper[4906]: I1123 08:56:44.821050 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgk8q"] Nov 23 08:56:46 crc kubenswrapper[4906]: I1123 08:56:46.732001 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tgk8q" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="registry-server" containerID="cri-o://7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c" gracePeriod=2 Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.094470 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.113177 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-catalog-content\") pod \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.131880 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" (UID: "c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.215045 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-utilities\") pod \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.215146 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc4ml\" (UniqueName: \"kubernetes.io/projected/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-kube-api-access-jc4ml\") pod \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\" (UID: \"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a\") " Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.215597 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.215879 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-utilities" (OuterVolumeSpecName: "utilities") pod "c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" (UID: "c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.219977 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-kube-api-access-jc4ml" (OuterVolumeSpecName: "kube-api-access-jc4ml") pod "c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" (UID: "c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a"). InnerVolumeSpecName "kube-api-access-jc4ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.317352 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.317392 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc4ml\" (UniqueName: \"kubernetes.io/projected/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a-kube-api-access-jc4ml\") on node \"crc\" DevicePath \"\"" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.741499 4906 generic.go:334] "Generic (PLEG): container finished" podID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerID="7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c" exitCode=0 Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.741545 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgk8q" event={"ID":"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a","Type":"ContainerDied","Data":"7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c"} Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.741576 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgk8q" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.741590 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgk8q" event={"ID":"c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a","Type":"ContainerDied","Data":"2be151dfe3bcfa254c5172c93110c70715b3f3e7fd1fc2210cf4e9c58f874f86"} Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.741611 4906 scope.go:117] "RemoveContainer" containerID="7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.762836 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgk8q"] Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.764256 4906 scope.go:117] "RemoveContainer" containerID="247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.770325 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgk8q"] Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.786095 4906 scope.go:117] "RemoveContainer" containerID="dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.808415 4906 scope.go:117] "RemoveContainer" containerID="7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c" Nov 23 08:56:47 crc kubenswrapper[4906]: E1123 08:56:47.809091 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c\": container with ID starting with 7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c not found: ID does not exist" containerID="7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.809137 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c"} err="failed to get container status \"7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c\": rpc error: code = NotFound desc = could not find container \"7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c\": container with ID starting with 7da8e70a5948e8cf2fc407bd5d36ade6005f55a835d28052946610df0145fe1c not found: ID does not exist" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.809164 4906 scope.go:117] "RemoveContainer" containerID="247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4" Nov 23 08:56:47 crc kubenswrapper[4906]: E1123 08:56:47.809610 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4\": container with ID starting with 247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4 not found: ID does not exist" containerID="247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.809661 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4"} err="failed to get container status \"247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4\": rpc error: code = NotFound desc = could not find container \"247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4\": container with ID starting with 247eb645af4687e3b3894f4afb48f4576fe811ef006a3acd3d6fe9e1093ca2e4 not found: ID does not exist" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.809724 4906 scope.go:117] "RemoveContainer" containerID="dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6" Nov 23 08:56:47 crc kubenswrapper[4906]: E1123 08:56:47.810084 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6\": container with ID starting with dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6 not found: ID does not exist" containerID="dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6" Nov 23 08:56:47 crc kubenswrapper[4906]: I1123 08:56:47.810108 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6"} err="failed to get container status \"dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6\": rpc error: code = NotFound desc = could not find container \"dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6\": container with ID starting with dcc06ff77cdfc97ffb096615ecf5b1e995e5e6bc953b51d69a381f7904203ba6 not found: ID does not exist" Nov 23 08:56:48 crc kubenswrapper[4906]: I1123 08:56:48.356983 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:56:48 crc kubenswrapper[4906]: E1123 08:56:48.357273 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:56:49 crc kubenswrapper[4906]: I1123 08:56:49.365896 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" path="/var/lib/kubelet/pods/c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a/volumes" Nov 23 08:57:02 crc kubenswrapper[4906]: I1123 08:57:02.356320 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:57:02 crc kubenswrapper[4906]: E1123 08:57:02.357027 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:57:13 crc kubenswrapper[4906]: I1123 08:57:13.357318 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:57:13 crc kubenswrapper[4906]: E1123 08:57:13.358445 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:57:25 crc kubenswrapper[4906]: I1123 08:57:25.371024 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:57:25 crc kubenswrapper[4906]: E1123 08:57:25.371980 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:57:38 crc kubenswrapper[4906]: I1123 08:57:38.357069 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:57:38 crc kubenswrapper[4906]: E1123 08:57:38.357875 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:57:53 crc kubenswrapper[4906]: I1123 08:57:53.356791 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:57:53 crc kubenswrapper[4906]: E1123 08:57:53.357646 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:58:06 crc kubenswrapper[4906]: I1123 08:58:06.357092 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:58:06 crc kubenswrapper[4906]: E1123 08:58:06.357943 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:58:21 crc kubenswrapper[4906]: I1123 08:58:21.357413 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:58:21 crc kubenswrapper[4906]: E1123 08:58:21.358647 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:58:33 crc kubenswrapper[4906]: I1123 08:58:33.356888 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:58:33 crc kubenswrapper[4906]: E1123 08:58:33.357863 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:58:45 crc kubenswrapper[4906]: I1123 08:58:45.361526 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:58:45 crc kubenswrapper[4906]: E1123 08:58:45.363473 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 08:59:00 crc kubenswrapper[4906]: I1123 08:59:00.356137 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 08:59:00 crc kubenswrapper[4906]: I1123 08:59:00.947832 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"1ca4e815ca14b2f4bbbb9d728368954e440202703782310721414a99e206ec27"} Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.146476 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v"] Nov 23 09:00:00 crc kubenswrapper[4906]: E1123 09:00:00.147243 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="extract-content" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.147273 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="extract-content" Nov 23 09:00:00 crc kubenswrapper[4906]: E1123 09:00:00.147302 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="extract-utilities" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.147308 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="extract-utilities" Nov 23 09:00:00 crc kubenswrapper[4906]: E1123 09:00:00.147318 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="registry-server" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.147324 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="registry-server" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.147457 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5e74917-b7b3-4983-b8aa-7dc9f5e17e0a" containerName="registry-server" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.148167 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.150703 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.150865 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.163771 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v"] Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.265234 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/42177685-7754-400a-9394-f6298e1e81fa-config-volume\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.265468 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/42177685-7754-400a-9394-f6298e1e81fa-secret-volume\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.265570 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvqc5\" (UniqueName: \"kubernetes.io/projected/42177685-7754-400a-9394-f6298e1e81fa-kube-api-access-rvqc5\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.366688 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvqc5\" (UniqueName: \"kubernetes.io/projected/42177685-7754-400a-9394-f6298e1e81fa-kube-api-access-rvqc5\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.366807 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/42177685-7754-400a-9394-f6298e1e81fa-config-volume\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.366851 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/42177685-7754-400a-9394-f6298e1e81fa-secret-volume\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.368664 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/42177685-7754-400a-9394-f6298e1e81fa-config-volume\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.375049 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/42177685-7754-400a-9394-f6298e1e81fa-secret-volume\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.382957 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvqc5\" (UniqueName: \"kubernetes.io/projected/42177685-7754-400a-9394-f6298e1e81fa-kube-api-access-rvqc5\") pod \"collect-profiles-29398140-4sz6v\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.524500 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:00 crc kubenswrapper[4906]: I1123 09:00:00.935195 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v"] Nov 23 09:00:01 crc kubenswrapper[4906]: I1123 09:00:01.418489 4906 generic.go:334] "Generic (PLEG): container finished" podID="42177685-7754-400a-9394-f6298e1e81fa" containerID="23d274e7e567d60fc365418886eae22cc83825a7bd93f8474eecfa872b9fdce6" exitCode=0 Nov 23 09:00:01 crc kubenswrapper[4906]: I1123 09:00:01.418599 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" event={"ID":"42177685-7754-400a-9394-f6298e1e81fa","Type":"ContainerDied","Data":"23d274e7e567d60fc365418886eae22cc83825a7bd93f8474eecfa872b9fdce6"} Nov 23 09:00:01 crc kubenswrapper[4906]: I1123 09:00:01.419643 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" event={"ID":"42177685-7754-400a-9394-f6298e1e81fa","Type":"ContainerStarted","Data":"117c12eccc178a38d0380aa8fd54d783e03c200e3a89c47159193c547abb4a99"} Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.772138 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.810473 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/42177685-7754-400a-9394-f6298e1e81fa-config-volume\") pod \"42177685-7754-400a-9394-f6298e1e81fa\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.810883 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/42177685-7754-400a-9394-f6298e1e81fa-secret-volume\") pod \"42177685-7754-400a-9394-f6298e1e81fa\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.810954 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvqc5\" (UniqueName: \"kubernetes.io/projected/42177685-7754-400a-9394-f6298e1e81fa-kube-api-access-rvqc5\") pod \"42177685-7754-400a-9394-f6298e1e81fa\" (UID: \"42177685-7754-400a-9394-f6298e1e81fa\") " Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.812387 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42177685-7754-400a-9394-f6298e1e81fa-config-volume" (OuterVolumeSpecName: "config-volume") pod "42177685-7754-400a-9394-f6298e1e81fa" (UID: "42177685-7754-400a-9394-f6298e1e81fa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.818218 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42177685-7754-400a-9394-f6298e1e81fa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "42177685-7754-400a-9394-f6298e1e81fa" (UID: "42177685-7754-400a-9394-f6298e1e81fa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.823956 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42177685-7754-400a-9394-f6298e1e81fa-kube-api-access-rvqc5" (OuterVolumeSpecName: "kube-api-access-rvqc5") pod "42177685-7754-400a-9394-f6298e1e81fa" (UID: "42177685-7754-400a-9394-f6298e1e81fa"). InnerVolumeSpecName "kube-api-access-rvqc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.913270 4906 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/42177685-7754-400a-9394-f6298e1e81fa-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.913757 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvqc5\" (UniqueName: \"kubernetes.io/projected/42177685-7754-400a-9394-f6298e1e81fa-kube-api-access-rvqc5\") on node \"crc\" DevicePath \"\"" Nov 23 09:00:02 crc kubenswrapper[4906]: I1123 09:00:02.913932 4906 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/42177685-7754-400a-9394-f6298e1e81fa-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 09:00:03 crc kubenswrapper[4906]: I1123 09:00:03.451591 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" event={"ID":"42177685-7754-400a-9394-f6298e1e81fa","Type":"ContainerDied","Data":"117c12eccc178a38d0380aa8fd54d783e03c200e3a89c47159193c547abb4a99"} Nov 23 09:00:03 crc kubenswrapper[4906]: I1123 09:00:03.451658 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="117c12eccc178a38d0380aa8fd54d783e03c200e3a89c47159193c547abb4a99" Nov 23 09:00:03 crc kubenswrapper[4906]: I1123 09:00:03.451836 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398140-4sz6v" Nov 23 09:00:03 crc kubenswrapper[4906]: I1123 09:00:03.900832 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45"] Nov 23 09:00:03 crc kubenswrapper[4906]: I1123 09:00:03.904734 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398095-86c45"] Nov 23 09:00:05 crc kubenswrapper[4906]: I1123 09:00:05.370797 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6efcfd7-e3ef-404f-9a43-8cb8514e43b8" path="/var/lib/kubelet/pods/f6efcfd7-e3ef-404f-9a43-8cb8514e43b8/volumes" Nov 23 09:00:59 crc kubenswrapper[4906]: I1123 09:00:59.485187 4906 scope.go:117] "RemoveContainer" containerID="fb87792ee365d7e810b540eee0d4886bfa04b51e9f9694dcfa762128371b778d" Nov 23 09:01:20 crc kubenswrapper[4906]: I1123 09:01:20.945345 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:01:20 crc kubenswrapper[4906]: I1123 09:01:20.946073 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 09:01:38 crc kubenswrapper[4906]: E1123 09:01:38.040196 4906 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Nov 23 09:01:50 crc kubenswrapper[4906]: I1123 09:01:50.946520 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:01:50 crc kubenswrapper[4906]: I1123 09:01:50.947168 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 09:02:20 crc kubenswrapper[4906]: I1123 09:02:20.945617 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:02:20 crc kubenswrapper[4906]: I1123 09:02:20.946079 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 09:02:20 crc kubenswrapper[4906]: I1123 09:02:20.946123 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 09:02:20 crc kubenswrapper[4906]: I1123 09:02:20.946700 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ca4e815ca14b2f4bbbb9d728368954e440202703782310721414a99e206ec27"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 09:02:20 crc kubenswrapper[4906]: I1123 09:02:20.946750 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://1ca4e815ca14b2f4bbbb9d728368954e440202703782310721414a99e206ec27" gracePeriod=600 Nov 23 09:02:21 crc kubenswrapper[4906]: I1123 09:02:21.645472 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="1ca4e815ca14b2f4bbbb9d728368954e440202703782310721414a99e206ec27" exitCode=0 Nov 23 09:02:21 crc kubenswrapper[4906]: I1123 09:02:21.645558 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"1ca4e815ca14b2f4bbbb9d728368954e440202703782310721414a99e206ec27"} Nov 23 09:02:21 crc kubenswrapper[4906]: I1123 09:02:21.646069 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357"} Nov 23 09:02:21 crc kubenswrapper[4906]: I1123 09:02:21.646088 4906 scope.go:117] "RemoveContainer" containerID="c511ef68593b91aade7ba8111b9b6c510d31abccfdae12c488f3c9d51c9b7f96" Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.893751 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9jrnr"] Nov 23 09:02:40 crc kubenswrapper[4906]: E1123 09:02:40.894578 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42177685-7754-400a-9394-f6298e1e81fa" containerName="collect-profiles" Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.894596 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="42177685-7754-400a-9394-f6298e1e81fa" containerName="collect-profiles" Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.894827 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="42177685-7754-400a-9394-f6298e1e81fa" containerName="collect-profiles" Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.895899 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.906156 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9jrnr"] Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.923612 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-catalog-content\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.924107 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-utilities\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:40 crc kubenswrapper[4906]: I1123 09:02:40.925610 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzwf2\" (UniqueName: \"kubernetes.io/projected/c48ec8ee-f249-466c-b876-8e0382396ff5-kube-api-access-wzwf2\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.027122 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzwf2\" (UniqueName: \"kubernetes.io/projected/c48ec8ee-f249-466c-b876-8e0382396ff5-kube-api-access-wzwf2\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.027190 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-catalog-content\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.027304 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-utilities\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.027792 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-catalog-content\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.027824 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-utilities\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.059900 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzwf2\" (UniqueName: \"kubernetes.io/projected/c48ec8ee-f249-466c-b876-8e0382396ff5-kube-api-access-wzwf2\") pod \"redhat-operators-9jrnr\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.226489 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.433259 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9jrnr"] Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.807921 4906 generic.go:334] "Generic (PLEG): container finished" podID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerID="0b73ba6aba85f26c1630e06093afbc7ff68c5e5e6ac54846a684b47870283fe3" exitCode=0 Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.807999 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jrnr" event={"ID":"c48ec8ee-f249-466c-b876-8e0382396ff5","Type":"ContainerDied","Data":"0b73ba6aba85f26c1630e06093afbc7ff68c5e5e6ac54846a684b47870283fe3"} Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.808269 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jrnr" event={"ID":"c48ec8ee-f249-466c-b876-8e0382396ff5","Type":"ContainerStarted","Data":"a10e150ee50956a813ec43331796f888bc67bef0b2b7c95ae6791ca1bbf35143"} Nov 23 09:02:41 crc kubenswrapper[4906]: I1123 09:02:41.809541 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 09:02:42 crc kubenswrapper[4906]: I1123 09:02:42.816554 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jrnr" event={"ID":"c48ec8ee-f249-466c-b876-8e0382396ff5","Type":"ContainerStarted","Data":"20d21f276adc3aa9951605fd5ae312815c48f4c1ed0c0a5aac5d26b6bfcbce1a"} Nov 23 09:02:43 crc kubenswrapper[4906]: I1123 09:02:43.826484 4906 generic.go:334] "Generic (PLEG): container finished" podID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerID="20d21f276adc3aa9951605fd5ae312815c48f4c1ed0c0a5aac5d26b6bfcbce1a" exitCode=0 Nov 23 09:02:43 crc kubenswrapper[4906]: I1123 09:02:43.826524 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jrnr" event={"ID":"c48ec8ee-f249-466c-b876-8e0382396ff5","Type":"ContainerDied","Data":"20d21f276adc3aa9951605fd5ae312815c48f4c1ed0c0a5aac5d26b6bfcbce1a"} Nov 23 09:02:44 crc kubenswrapper[4906]: I1123 09:02:44.837065 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jrnr" event={"ID":"c48ec8ee-f249-466c-b876-8e0382396ff5","Type":"ContainerStarted","Data":"5877fa62529d3573bcedf7a1b7d6435115a60b35aece559f66bc7ac8f457dd16"} Nov 23 09:02:44 crc kubenswrapper[4906]: I1123 09:02:44.869614 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9jrnr" podStartSLOduration=2.329738929 podStartE2EDuration="4.869582334s" podCreationTimestamp="2025-11-23 09:02:40 +0000 UTC" firstStartedPulling="2025-11-23 09:02:41.809274317 +0000 UTC m=+7977.322665620" lastFinishedPulling="2025-11-23 09:02:44.349117672 +0000 UTC m=+7979.862509025" observedRunningTime="2025-11-23 09:02:44.865826744 +0000 UTC m=+7980.379218067" watchObservedRunningTime="2025-11-23 09:02:44.869582334 +0000 UTC m=+7980.382973647" Nov 23 09:02:51 crc kubenswrapper[4906]: I1123 09:02:51.227617 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:51 crc kubenswrapper[4906]: I1123 09:02:51.227957 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:51 crc kubenswrapper[4906]: I1123 09:02:51.294575 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:51 crc kubenswrapper[4906]: I1123 09:02:51.955129 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:52 crc kubenswrapper[4906]: I1123 09:02:52.005581 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9jrnr"] Nov 23 09:02:53 crc kubenswrapper[4906]: I1123 09:02:53.909065 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9jrnr" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="registry-server" containerID="cri-o://5877fa62529d3573bcedf7a1b7d6435115a60b35aece559f66bc7ac8f457dd16" gracePeriod=2 Nov 23 09:02:54 crc kubenswrapper[4906]: I1123 09:02:54.918232 4906 generic.go:334] "Generic (PLEG): container finished" podID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerID="5877fa62529d3573bcedf7a1b7d6435115a60b35aece559f66bc7ac8f457dd16" exitCode=0 Nov 23 09:02:54 crc kubenswrapper[4906]: I1123 09:02:54.918465 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jrnr" event={"ID":"c48ec8ee-f249-466c-b876-8e0382396ff5","Type":"ContainerDied","Data":"5877fa62529d3573bcedf7a1b7d6435115a60b35aece559f66bc7ac8f457dd16"} Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.397060 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.549815 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzwf2\" (UniqueName: \"kubernetes.io/projected/c48ec8ee-f249-466c-b876-8e0382396ff5-kube-api-access-wzwf2\") pod \"c48ec8ee-f249-466c-b876-8e0382396ff5\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.549938 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-catalog-content\") pod \"c48ec8ee-f249-466c-b876-8e0382396ff5\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.550018 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-utilities\") pod \"c48ec8ee-f249-466c-b876-8e0382396ff5\" (UID: \"c48ec8ee-f249-466c-b876-8e0382396ff5\") " Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.551878 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-utilities" (OuterVolumeSpecName: "utilities") pod "c48ec8ee-f249-466c-b876-8e0382396ff5" (UID: "c48ec8ee-f249-466c-b876-8e0382396ff5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.554759 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48ec8ee-f249-466c-b876-8e0382396ff5-kube-api-access-wzwf2" (OuterVolumeSpecName: "kube-api-access-wzwf2") pod "c48ec8ee-f249-466c-b876-8e0382396ff5" (UID: "c48ec8ee-f249-466c-b876-8e0382396ff5"). InnerVolumeSpecName "kube-api-access-wzwf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.652522 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.652557 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzwf2\" (UniqueName: \"kubernetes.io/projected/c48ec8ee-f249-466c-b876-8e0382396ff5-kube-api-access-wzwf2\") on node \"crc\" DevicePath \"\"" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.652637 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c48ec8ee-f249-466c-b876-8e0382396ff5" (UID: "c48ec8ee-f249-466c-b876-8e0382396ff5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.754149 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c48ec8ee-f249-466c-b876-8e0382396ff5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.929426 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jrnr" event={"ID":"c48ec8ee-f249-466c-b876-8e0382396ff5","Type":"ContainerDied","Data":"a10e150ee50956a813ec43331796f888bc67bef0b2b7c95ae6791ca1bbf35143"} Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.929475 4906 scope.go:117] "RemoveContainer" containerID="5877fa62529d3573bcedf7a1b7d6435115a60b35aece559f66bc7ac8f457dd16" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.929596 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jrnr" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.962370 4906 scope.go:117] "RemoveContainer" containerID="20d21f276adc3aa9951605fd5ae312815c48f4c1ed0c0a5aac5d26b6bfcbce1a" Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.967497 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9jrnr"] Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.971772 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9jrnr"] Nov 23 09:02:55 crc kubenswrapper[4906]: I1123 09:02:55.984372 4906 scope.go:117] "RemoveContainer" containerID="0b73ba6aba85f26c1630e06093afbc7ff68c5e5e6ac54846a684b47870283fe3" Nov 23 09:02:57 crc kubenswrapper[4906]: I1123 09:02:57.372888 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" path="/var/lib/kubelet/pods/c48ec8ee-f249-466c-b876-8e0382396ff5/volumes" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.454921 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dtbv8"] Nov 23 09:03:03 crc kubenswrapper[4906]: E1123 09:03:03.456559 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="extract-utilities" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.456585 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="extract-utilities" Nov 23 09:03:03 crc kubenswrapper[4906]: E1123 09:03:03.456609 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="registry-server" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.456621 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="registry-server" Nov 23 09:03:03 crc kubenswrapper[4906]: E1123 09:03:03.456639 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="extract-content" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.456649 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="extract-content" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.456901 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48ec8ee-f249-466c-b876-8e0382396ff5" containerName="registry-server" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.458960 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.459514 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtbv8"] Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.506052 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-utilities\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.506101 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-catalog-content\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.506131 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqmm4\" (UniqueName: \"kubernetes.io/projected/be78c0d3-dd76-4d54-81e0-e670306098ef-kube-api-access-lqmm4\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.607864 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-utilities\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.607953 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-catalog-content\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.607990 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqmm4\" (UniqueName: \"kubernetes.io/projected/be78c0d3-dd76-4d54-81e0-e670306098ef-kube-api-access-lqmm4\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.608439 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-utilities\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.608498 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-catalog-content\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.630747 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqmm4\" (UniqueName: \"kubernetes.io/projected/be78c0d3-dd76-4d54-81e0-e670306098ef-kube-api-access-lqmm4\") pod \"certified-operators-dtbv8\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:03 crc kubenswrapper[4906]: I1123 09:03:03.786244 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:04 crc kubenswrapper[4906]: I1123 09:03:04.330969 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtbv8"] Nov 23 09:03:05 crc kubenswrapper[4906]: I1123 09:03:05.013057 4906 generic.go:334] "Generic (PLEG): container finished" podID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerID="431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00" exitCode=0 Nov 23 09:03:05 crc kubenswrapper[4906]: I1123 09:03:05.013099 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtbv8" event={"ID":"be78c0d3-dd76-4d54-81e0-e670306098ef","Type":"ContainerDied","Data":"431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00"} Nov 23 09:03:05 crc kubenswrapper[4906]: I1123 09:03:05.013128 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtbv8" event={"ID":"be78c0d3-dd76-4d54-81e0-e670306098ef","Type":"ContainerStarted","Data":"04e90aad290b257c0127ed6be81ee3d7e38ad388a44380b09e70334803b03ebb"} Nov 23 09:03:06 crc kubenswrapper[4906]: I1123 09:03:06.024118 4906 generic.go:334] "Generic (PLEG): container finished" podID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerID="70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08" exitCode=0 Nov 23 09:03:06 crc kubenswrapper[4906]: I1123 09:03:06.024331 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtbv8" event={"ID":"be78c0d3-dd76-4d54-81e0-e670306098ef","Type":"ContainerDied","Data":"70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08"} Nov 23 09:03:07 crc kubenswrapper[4906]: I1123 09:03:07.036260 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtbv8" event={"ID":"be78c0d3-dd76-4d54-81e0-e670306098ef","Type":"ContainerStarted","Data":"378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7"} Nov 23 09:03:07 crc kubenswrapper[4906]: I1123 09:03:07.055763 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dtbv8" podStartSLOduration=2.6598446190000002 podStartE2EDuration="4.055748401s" podCreationTimestamp="2025-11-23 09:03:03 +0000 UTC" firstStartedPulling="2025-11-23 09:03:05.016324604 +0000 UTC m=+8000.529715937" lastFinishedPulling="2025-11-23 09:03:06.412228376 +0000 UTC m=+8001.925619719" observedRunningTime="2025-11-23 09:03:07.053912501 +0000 UTC m=+8002.567303824" watchObservedRunningTime="2025-11-23 09:03:07.055748401 +0000 UTC m=+8002.569139704" Nov 23 09:03:13 crc kubenswrapper[4906]: I1123 09:03:13.787202 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:13 crc kubenswrapper[4906]: I1123 09:03:13.787807 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:13 crc kubenswrapper[4906]: I1123 09:03:13.857894 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:14 crc kubenswrapper[4906]: I1123 09:03:14.144146 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:14 crc kubenswrapper[4906]: I1123 09:03:14.202194 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtbv8"] Nov 23 09:03:16 crc kubenswrapper[4906]: I1123 09:03:16.114846 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dtbv8" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="registry-server" containerID="cri-o://378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7" gracePeriod=2 Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.022711 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.122004 4906 generic.go:334] "Generic (PLEG): container finished" podID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerID="378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7" exitCode=0 Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.122046 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtbv8" event={"ID":"be78c0d3-dd76-4d54-81e0-e670306098ef","Type":"ContainerDied","Data":"378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7"} Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.122074 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtbv8" event={"ID":"be78c0d3-dd76-4d54-81e0-e670306098ef","Type":"ContainerDied","Data":"04e90aad290b257c0127ed6be81ee3d7e38ad388a44380b09e70334803b03ebb"} Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.122083 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtbv8" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.122103 4906 scope.go:117] "RemoveContainer" containerID="378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.136036 4906 scope.go:117] "RemoveContainer" containerID="70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.149034 4906 scope.go:117] "RemoveContainer" containerID="431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.156642 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-utilities\") pod \"be78c0d3-dd76-4d54-81e0-e670306098ef\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.156739 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqmm4\" (UniqueName: \"kubernetes.io/projected/be78c0d3-dd76-4d54-81e0-e670306098ef-kube-api-access-lqmm4\") pod \"be78c0d3-dd76-4d54-81e0-e670306098ef\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.156807 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-catalog-content\") pod \"be78c0d3-dd76-4d54-81e0-e670306098ef\" (UID: \"be78c0d3-dd76-4d54-81e0-e670306098ef\") " Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.157799 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-utilities" (OuterVolumeSpecName: "utilities") pod "be78c0d3-dd76-4d54-81e0-e670306098ef" (UID: "be78c0d3-dd76-4d54-81e0-e670306098ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.163629 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be78c0d3-dd76-4d54-81e0-e670306098ef-kube-api-access-lqmm4" (OuterVolumeSpecName: "kube-api-access-lqmm4") pod "be78c0d3-dd76-4d54-81e0-e670306098ef" (UID: "be78c0d3-dd76-4d54-81e0-e670306098ef"). InnerVolumeSpecName "kube-api-access-lqmm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.197842 4906 scope.go:117] "RemoveContainer" containerID="378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7" Nov 23 09:03:17 crc kubenswrapper[4906]: E1123 09:03:17.198274 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7\": container with ID starting with 378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7 not found: ID does not exist" containerID="378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.198303 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7"} err="failed to get container status \"378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7\": rpc error: code = NotFound desc = could not find container \"378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7\": container with ID starting with 378fb49654c7aa1e0a5d39195695f6b67bf800f6c64c113e0eb5093fd8c9d5b7 not found: ID does not exist" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.198323 4906 scope.go:117] "RemoveContainer" containerID="70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08" Nov 23 09:03:17 crc kubenswrapper[4906]: E1123 09:03:17.198628 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08\": container with ID starting with 70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08 not found: ID does not exist" containerID="70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.198652 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08"} err="failed to get container status \"70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08\": rpc error: code = NotFound desc = could not find container \"70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08\": container with ID starting with 70ba9133157c98ff85284bbff6bdc5e1133b4e17ceeafebcb19447062d290d08 not found: ID does not exist" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.198666 4906 scope.go:117] "RemoveContainer" containerID="431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00" Nov 23 09:03:17 crc kubenswrapper[4906]: E1123 09:03:17.198939 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00\": container with ID starting with 431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00 not found: ID does not exist" containerID="431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.198989 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00"} err="failed to get container status \"431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00\": rpc error: code = NotFound desc = could not find container \"431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00\": container with ID starting with 431cdfe9216d497c4791583c9c889a3eeef04bc0b2b4033beda2b3ac6e506a00 not found: ID does not exist" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.213074 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be78c0d3-dd76-4d54-81e0-e670306098ef" (UID: "be78c0d3-dd76-4d54-81e0-e670306098ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.258016 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqmm4\" (UniqueName: \"kubernetes.io/projected/be78c0d3-dd76-4d54-81e0-e670306098ef-kube-api-access-lqmm4\") on node \"crc\" DevicePath \"\"" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.258057 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.258072 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be78c0d3-dd76-4d54-81e0-e670306098ef-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.448968 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtbv8"] Nov 23 09:03:17 crc kubenswrapper[4906]: I1123 09:03:17.454483 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dtbv8"] Nov 23 09:03:19 crc kubenswrapper[4906]: I1123 09:03:19.374317 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" path="/var/lib/kubelet/pods/be78c0d3-dd76-4d54-81e0-e670306098ef/volumes" Nov 23 09:04:50 crc kubenswrapper[4906]: I1123 09:04:50.946079 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:04:50 crc kubenswrapper[4906]: I1123 09:04:50.946736 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 09:05:20 crc kubenswrapper[4906]: I1123 09:05:20.946512 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:05:20 crc kubenswrapper[4906]: I1123 09:05:20.947229 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 09:05:50 crc kubenswrapper[4906]: I1123 09:05:50.946160 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:05:50 crc kubenswrapper[4906]: I1123 09:05:50.946879 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 09:05:50 crc kubenswrapper[4906]: I1123 09:05:50.946943 4906 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" Nov 23 09:05:50 crc kubenswrapper[4906]: I1123 09:05:50.947921 4906 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357"} pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 09:05:50 crc kubenswrapper[4906]: I1123 09:05:50.948023 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" containerID="cri-o://e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" gracePeriod=600 Nov 23 09:05:51 crc kubenswrapper[4906]: E1123 09:05:51.080860 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:05:51 crc kubenswrapper[4906]: I1123 09:05:51.559271 4906 generic.go:334] "Generic (PLEG): container finished" podID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" exitCode=0 Nov 23 09:05:51 crc kubenswrapper[4906]: I1123 09:05:51.559393 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerDied","Data":"e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357"} Nov 23 09:05:51 crc kubenswrapper[4906]: I1123 09:05:51.559556 4906 scope.go:117] "RemoveContainer" containerID="1ca4e815ca14b2f4bbbb9d728368954e440202703782310721414a99e206ec27" Nov 23 09:05:51 crc kubenswrapper[4906]: I1123 09:05:51.560996 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:05:51 crc kubenswrapper[4906]: E1123 09:05:51.561755 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:06:04 crc kubenswrapper[4906]: I1123 09:06:04.356532 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:06:04 crc kubenswrapper[4906]: E1123 09:06:04.357261 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:06:15 crc kubenswrapper[4906]: I1123 09:06:15.360457 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:06:15 crc kubenswrapper[4906]: E1123 09:06:15.361180 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:06:29 crc kubenswrapper[4906]: I1123 09:06:29.356787 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:06:29 crc kubenswrapper[4906]: E1123 09:06:29.357455 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:06:40 crc kubenswrapper[4906]: I1123 09:06:40.356576 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:06:40 crc kubenswrapper[4906]: E1123 09:06:40.357674 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:06:51 crc kubenswrapper[4906]: I1123 09:06:51.986571 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8mwzf"] Nov 23 09:06:51 crc kubenswrapper[4906]: E1123 09:06:51.987927 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="registry-server" Nov 23 09:06:51 crc kubenswrapper[4906]: I1123 09:06:51.987956 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="registry-server" Nov 23 09:06:51 crc kubenswrapper[4906]: E1123 09:06:51.987985 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="extract-content" Nov 23 09:06:51 crc kubenswrapper[4906]: I1123 09:06:51.988001 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="extract-content" Nov 23 09:06:51 crc kubenswrapper[4906]: E1123 09:06:51.988040 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="extract-utilities" Nov 23 09:06:51 crc kubenswrapper[4906]: I1123 09:06:51.988053 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="extract-utilities" Nov 23 09:06:51 crc kubenswrapper[4906]: I1123 09:06:51.988387 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="be78c0d3-dd76-4d54-81e0-e670306098ef" containerName="registry-server" Nov 23 09:06:51 crc kubenswrapper[4906]: I1123 09:06:51.990272 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.005794 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8mwzf"] Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.126918 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-catalog-content\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.127063 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4772f\" (UniqueName: \"kubernetes.io/projected/601c8761-1e89-4278-ad44-fd1b534d39b4-kube-api-access-4772f\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.127111 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-utilities\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.228739 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-catalog-content\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.228796 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-catalog-content\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.228901 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4772f\" (UniqueName: \"kubernetes.io/projected/601c8761-1e89-4278-ad44-fd1b534d39b4-kube-api-access-4772f\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.229768 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-utilities\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.229360 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-utilities\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.257520 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4772f\" (UniqueName: \"kubernetes.io/projected/601c8761-1e89-4278-ad44-fd1b534d39b4-kube-api-access-4772f\") pod \"community-operators-8mwzf\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.313418 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:06:52 crc kubenswrapper[4906]: I1123 09:06:52.807620 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8mwzf"] Nov 23 09:06:53 crc kubenswrapper[4906]: I1123 09:06:53.102117 4906 generic.go:334] "Generic (PLEG): container finished" podID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerID="eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e" exitCode=0 Nov 23 09:06:53 crc kubenswrapper[4906]: I1123 09:06:53.102196 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mwzf" event={"ID":"601c8761-1e89-4278-ad44-fd1b534d39b4","Type":"ContainerDied","Data":"eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e"} Nov 23 09:06:53 crc kubenswrapper[4906]: I1123 09:06:53.102396 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mwzf" event={"ID":"601c8761-1e89-4278-ad44-fd1b534d39b4","Type":"ContainerStarted","Data":"0152dbacfc0b7528747a650860a1ae68eff786906e3e0dcd815997ffa4af9ec1"} Nov 23 09:06:53 crc kubenswrapper[4906]: I1123 09:06:53.357147 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:06:53 crc kubenswrapper[4906]: E1123 09:06:53.357367 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:06:54 crc kubenswrapper[4906]: I1123 09:06:54.111346 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mwzf" event={"ID":"601c8761-1e89-4278-ad44-fd1b534d39b4","Type":"ContainerStarted","Data":"e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203"} Nov 23 09:06:55 crc kubenswrapper[4906]: I1123 09:06:55.122943 4906 generic.go:334] "Generic (PLEG): container finished" podID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerID="e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203" exitCode=0 Nov 23 09:06:55 crc kubenswrapper[4906]: I1123 09:06:55.123000 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mwzf" event={"ID":"601c8761-1e89-4278-ad44-fd1b534d39b4","Type":"ContainerDied","Data":"e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203"} Nov 23 09:06:56 crc kubenswrapper[4906]: I1123 09:06:56.135798 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mwzf" event={"ID":"601c8761-1e89-4278-ad44-fd1b534d39b4","Type":"ContainerStarted","Data":"d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c"} Nov 23 09:06:56 crc kubenswrapper[4906]: I1123 09:06:56.168524 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8mwzf" podStartSLOduration=2.698849137 podStartE2EDuration="5.16850378s" podCreationTimestamp="2025-11-23 09:06:51 +0000 UTC" firstStartedPulling="2025-11-23 09:06:53.10656387 +0000 UTC m=+8228.619955173" lastFinishedPulling="2025-11-23 09:06:55.576218523 +0000 UTC m=+8231.089609816" observedRunningTime="2025-11-23 09:06:56.163002264 +0000 UTC m=+8231.676393607" watchObservedRunningTime="2025-11-23 09:06:56.16850378 +0000 UTC m=+8231.681895093" Nov 23 09:07:02 crc kubenswrapper[4906]: I1123 09:07:02.313958 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:07:02 crc kubenswrapper[4906]: I1123 09:07:02.314518 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:07:02 crc kubenswrapper[4906]: I1123 09:07:02.398066 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:07:03 crc kubenswrapper[4906]: I1123 09:07:03.241568 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:07:03 crc kubenswrapper[4906]: I1123 09:07:03.288047 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8mwzf"] Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.217563 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8mwzf" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="registry-server" containerID="cri-o://d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c" gracePeriod=2 Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.672150 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.734597 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-utilities\") pod \"601c8761-1e89-4278-ad44-fd1b534d39b4\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.734671 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-catalog-content\") pod \"601c8761-1e89-4278-ad44-fd1b534d39b4\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.734716 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4772f\" (UniqueName: \"kubernetes.io/projected/601c8761-1e89-4278-ad44-fd1b534d39b4-kube-api-access-4772f\") pod \"601c8761-1e89-4278-ad44-fd1b534d39b4\" (UID: \"601c8761-1e89-4278-ad44-fd1b534d39b4\") " Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.735930 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-utilities" (OuterVolumeSpecName: "utilities") pod "601c8761-1e89-4278-ad44-fd1b534d39b4" (UID: "601c8761-1e89-4278-ad44-fd1b534d39b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.739784 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/601c8761-1e89-4278-ad44-fd1b534d39b4-kube-api-access-4772f" (OuterVolumeSpecName: "kube-api-access-4772f") pod "601c8761-1e89-4278-ad44-fd1b534d39b4" (UID: "601c8761-1e89-4278-ad44-fd1b534d39b4"). InnerVolumeSpecName "kube-api-access-4772f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.798222 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "601c8761-1e89-4278-ad44-fd1b534d39b4" (UID: "601c8761-1e89-4278-ad44-fd1b534d39b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.836261 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.836295 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/601c8761-1e89-4278-ad44-fd1b534d39b4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 09:07:05 crc kubenswrapper[4906]: I1123 09:07:05.836306 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4772f\" (UniqueName: \"kubernetes.io/projected/601c8761-1e89-4278-ad44-fd1b534d39b4-kube-api-access-4772f\") on node \"crc\" DevicePath \"\"" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.230132 4906 generic.go:334] "Generic (PLEG): container finished" podID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerID="d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c" exitCode=0 Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.230181 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mwzf" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.230201 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mwzf" event={"ID":"601c8761-1e89-4278-ad44-fd1b534d39b4","Type":"ContainerDied","Data":"d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c"} Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.230254 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mwzf" event={"ID":"601c8761-1e89-4278-ad44-fd1b534d39b4","Type":"ContainerDied","Data":"0152dbacfc0b7528747a650860a1ae68eff786906e3e0dcd815997ffa4af9ec1"} Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.230288 4906 scope.go:117] "RemoveContainer" containerID="d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.269841 4906 scope.go:117] "RemoveContainer" containerID="e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.277320 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8mwzf"] Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.287404 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8mwzf"] Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.294045 4906 scope.go:117] "RemoveContainer" containerID="eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.332646 4906 scope.go:117] "RemoveContainer" containerID="d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c" Nov 23 09:07:06 crc kubenswrapper[4906]: E1123 09:07:06.333112 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c\": container with ID starting with d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c not found: ID does not exist" containerID="d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.333159 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c"} err="failed to get container status \"d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c\": rpc error: code = NotFound desc = could not find container \"d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c\": container with ID starting with d2a6e5cba3c407a071a46050754d4a97c3b6519f2db268164e876cf6632cf43c not found: ID does not exist" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.333192 4906 scope.go:117] "RemoveContainer" containerID="e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203" Nov 23 09:07:06 crc kubenswrapper[4906]: E1123 09:07:06.333501 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203\": container with ID starting with e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203 not found: ID does not exist" containerID="e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.333545 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203"} err="failed to get container status \"e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203\": rpc error: code = NotFound desc = could not find container \"e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203\": container with ID starting with e3812a9b24e4c9f7d16ac908b4bc5cb9728a4b9f50b3e15706356f3104017203 not found: ID does not exist" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.333570 4906 scope.go:117] "RemoveContainer" containerID="eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e" Nov 23 09:07:06 crc kubenswrapper[4906]: E1123 09:07:06.333973 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e\": container with ID starting with eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e not found: ID does not exist" containerID="eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e" Nov 23 09:07:06 crc kubenswrapper[4906]: I1123 09:07:06.334014 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e"} err="failed to get container status \"eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e\": rpc error: code = NotFound desc = could not find container \"eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e\": container with ID starting with eb8897008a795cee831d4f72616a04287cc73b01335cee18131b0654c3fd925e not found: ID does not exist" Nov 23 09:07:07 crc kubenswrapper[4906]: I1123 09:07:07.368530 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" path="/var/lib/kubelet/pods/601c8761-1e89-4278-ad44-fd1b534d39b4/volumes" Nov 23 09:07:08 crc kubenswrapper[4906]: I1123 09:07:08.356529 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:07:08 crc kubenswrapper[4906]: E1123 09:07:08.356780 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.622326 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jkh9j"] Nov 23 09:07:13 crc kubenswrapper[4906]: E1123 09:07:13.623467 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="registry-server" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.623491 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="registry-server" Nov 23 09:07:13 crc kubenswrapper[4906]: E1123 09:07:13.623523 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="extract-utilities" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.623536 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="extract-utilities" Nov 23 09:07:13 crc kubenswrapper[4906]: E1123 09:07:13.623554 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="extract-content" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.623569 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="extract-content" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.624221 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="601c8761-1e89-4278-ad44-fd1b534d39b4" containerName="registry-server" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.626116 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.646856 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkh9j"] Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.758670 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-utilities\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.758817 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlj5f\" (UniqueName: \"kubernetes.io/projected/d93ec0bd-718f-4812-99d0-85ab356fd9db-kube-api-access-rlj5f\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.759560 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-catalog-content\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.860822 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-catalog-content\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.860888 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-utilities\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.860930 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlj5f\" (UniqueName: \"kubernetes.io/projected/d93ec0bd-718f-4812-99d0-85ab356fd9db-kube-api-access-rlj5f\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.861385 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-catalog-content\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.861745 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-utilities\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.896002 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlj5f\" (UniqueName: \"kubernetes.io/projected/d93ec0bd-718f-4812-99d0-85ab356fd9db-kube-api-access-rlj5f\") pod \"redhat-marketplace-jkh9j\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:13 crc kubenswrapper[4906]: I1123 09:07:13.954301 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:14 crc kubenswrapper[4906]: I1123 09:07:14.408322 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkh9j"] Nov 23 09:07:15 crc kubenswrapper[4906]: I1123 09:07:15.317830 4906 generic.go:334] "Generic (PLEG): container finished" podID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerID="ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4" exitCode=0 Nov 23 09:07:15 crc kubenswrapper[4906]: I1123 09:07:15.317932 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkh9j" event={"ID":"d93ec0bd-718f-4812-99d0-85ab356fd9db","Type":"ContainerDied","Data":"ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4"} Nov 23 09:07:15 crc kubenswrapper[4906]: I1123 09:07:15.320744 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkh9j" event={"ID":"d93ec0bd-718f-4812-99d0-85ab356fd9db","Type":"ContainerStarted","Data":"0f797a7edaab2cb3d3c067e138faece9ae8c1a1094a1c9d1e1156ccb822f0268"} Nov 23 09:07:16 crc kubenswrapper[4906]: I1123 09:07:16.328273 4906 generic.go:334] "Generic (PLEG): container finished" podID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerID="94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d" exitCode=0 Nov 23 09:07:16 crc kubenswrapper[4906]: I1123 09:07:16.328341 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkh9j" event={"ID":"d93ec0bd-718f-4812-99d0-85ab356fd9db","Type":"ContainerDied","Data":"94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d"} Nov 23 09:07:17 crc kubenswrapper[4906]: I1123 09:07:17.340404 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkh9j" event={"ID":"d93ec0bd-718f-4812-99d0-85ab356fd9db","Type":"ContainerStarted","Data":"dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12"} Nov 23 09:07:17 crc kubenswrapper[4906]: I1123 09:07:17.369444 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jkh9j" podStartSLOduration=2.960609496 podStartE2EDuration="4.369424932s" podCreationTimestamp="2025-11-23 09:07:13 +0000 UTC" firstStartedPulling="2025-11-23 09:07:15.319721521 +0000 UTC m=+8250.833112824" lastFinishedPulling="2025-11-23 09:07:16.728536947 +0000 UTC m=+8252.241928260" observedRunningTime="2025-11-23 09:07:17.365963029 +0000 UTC m=+8252.879354322" watchObservedRunningTime="2025-11-23 09:07:17.369424932 +0000 UTC m=+8252.882816245" Nov 23 09:07:22 crc kubenswrapper[4906]: I1123 09:07:22.356711 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:07:22 crc kubenswrapper[4906]: E1123 09:07:22.357413 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:07:23 crc kubenswrapper[4906]: I1123 09:07:23.955464 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:23 crc kubenswrapper[4906]: I1123 09:07:23.955544 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:24 crc kubenswrapper[4906]: I1123 09:07:24.013563 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:24 crc kubenswrapper[4906]: I1123 09:07:24.494662 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:24 crc kubenswrapper[4906]: I1123 09:07:24.552656 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkh9j"] Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.025145 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qbkj6/must-gather-nqcgv"] Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.026726 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.028911 4906 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qbkj6"/"default-dockercfg-pqfps" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.030042 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qbkj6"/"openshift-service-ca.crt" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.031647 4906 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qbkj6"/"kube-root-ca.crt" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.033495 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qbkj6/must-gather-nqcgv"] Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.044982 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc0ed17-e067-446e-9b4f-5027fd193264-must-gather-output\") pod \"must-gather-nqcgv\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.045044 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwcjh\" (UniqueName: \"kubernetes.io/projected/cdc0ed17-e067-446e-9b4f-5027fd193264-kube-api-access-rwcjh\") pod \"must-gather-nqcgv\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.146083 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc0ed17-e067-446e-9b4f-5027fd193264-must-gather-output\") pod \"must-gather-nqcgv\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.146145 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwcjh\" (UniqueName: \"kubernetes.io/projected/cdc0ed17-e067-446e-9b4f-5027fd193264-kube-api-access-rwcjh\") pod \"must-gather-nqcgv\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.146515 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc0ed17-e067-446e-9b4f-5027fd193264-must-gather-output\") pod \"must-gather-nqcgv\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.164715 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwcjh\" (UniqueName: \"kubernetes.io/projected/cdc0ed17-e067-446e-9b4f-5027fd193264-kube-api-access-rwcjh\") pod \"must-gather-nqcgv\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.342857 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.444495 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jkh9j" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="registry-server" containerID="cri-o://dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12" gracePeriod=2 Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.554207 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qbkj6/must-gather-nqcgv"] Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.823577 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.860943 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlj5f\" (UniqueName: \"kubernetes.io/projected/d93ec0bd-718f-4812-99d0-85ab356fd9db-kube-api-access-rlj5f\") pod \"d93ec0bd-718f-4812-99d0-85ab356fd9db\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.861138 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-utilities\") pod \"d93ec0bd-718f-4812-99d0-85ab356fd9db\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.861185 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-catalog-content\") pod \"d93ec0bd-718f-4812-99d0-85ab356fd9db\" (UID: \"d93ec0bd-718f-4812-99d0-85ab356fd9db\") " Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.879251 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-utilities" (OuterVolumeSpecName: "utilities") pod "d93ec0bd-718f-4812-99d0-85ab356fd9db" (UID: "d93ec0bd-718f-4812-99d0-85ab356fd9db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.879382 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93ec0bd-718f-4812-99d0-85ab356fd9db-kube-api-access-rlj5f" (OuterVolumeSpecName: "kube-api-access-rlj5f") pod "d93ec0bd-718f-4812-99d0-85ab356fd9db" (UID: "d93ec0bd-718f-4812-99d0-85ab356fd9db"). InnerVolumeSpecName "kube-api-access-rlj5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.880754 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d93ec0bd-718f-4812-99d0-85ab356fd9db" (UID: "d93ec0bd-718f-4812-99d0-85ab356fd9db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.962131 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.962174 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d93ec0bd-718f-4812-99d0-85ab356fd9db-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 09:07:26 crc kubenswrapper[4906]: I1123 09:07:26.962186 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlj5f\" (UniqueName: \"kubernetes.io/projected/d93ec0bd-718f-4812-99d0-85ab356fd9db-kube-api-access-rlj5f\") on node \"crc\" DevicePath \"\"" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.450440 4906 generic.go:334] "Generic (PLEG): container finished" podID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerID="dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12" exitCode=0 Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.450523 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkh9j" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.450526 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkh9j" event={"ID":"d93ec0bd-718f-4812-99d0-85ab356fd9db","Type":"ContainerDied","Data":"dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12"} Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.450592 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkh9j" event={"ID":"d93ec0bd-718f-4812-99d0-85ab356fd9db","Type":"ContainerDied","Data":"0f797a7edaab2cb3d3c067e138faece9ae8c1a1094a1c9d1e1156ccb822f0268"} Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.450623 4906 scope.go:117] "RemoveContainer" containerID="dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.452197 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" event={"ID":"cdc0ed17-e067-446e-9b4f-5027fd193264","Type":"ContainerStarted","Data":"fe095876145d8b8fd2ab0f8aa9ca0824afceddd5b8c5d6309496245f1ddbd084"} Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.493897 4906 scope.go:117] "RemoveContainer" containerID="94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.503750 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkh9j"] Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.518124 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkh9j"] Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.570143 4906 scope.go:117] "RemoveContainer" containerID="ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.592308 4906 scope.go:117] "RemoveContainer" containerID="dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12" Nov 23 09:07:27 crc kubenswrapper[4906]: E1123 09:07:27.592670 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12\": container with ID starting with dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12 not found: ID does not exist" containerID="dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.592708 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12"} err="failed to get container status \"dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12\": rpc error: code = NotFound desc = could not find container \"dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12\": container with ID starting with dac4ccd6fcb336c5c0e17e2afaff7af5938b7d32dc8abe38c84d792069967c12 not found: ID does not exist" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.592731 4906 scope.go:117] "RemoveContainer" containerID="94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d" Nov 23 09:07:27 crc kubenswrapper[4906]: E1123 09:07:27.592933 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d\": container with ID starting with 94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d not found: ID does not exist" containerID="94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.592954 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d"} err="failed to get container status \"94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d\": rpc error: code = NotFound desc = could not find container \"94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d\": container with ID starting with 94b5d9dd4360a791646de47eab31ada098d7f428dfad7148efc6c3f81511f42d not found: ID does not exist" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.592969 4906 scope.go:117] "RemoveContainer" containerID="ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4" Nov 23 09:07:27 crc kubenswrapper[4906]: E1123 09:07:27.593210 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4\": container with ID starting with ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4 not found: ID does not exist" containerID="ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4" Nov 23 09:07:27 crc kubenswrapper[4906]: I1123 09:07:27.593235 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4"} err="failed to get container status \"ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4\": rpc error: code = NotFound desc = could not find container \"ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4\": container with ID starting with ccba4dc76cb060a470013b80a351abfe391fe1a0e3b51039f678ed7f46079ad4 not found: ID does not exist" Nov 23 09:07:29 crc kubenswrapper[4906]: I1123 09:07:29.366449 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" path="/var/lib/kubelet/pods/d93ec0bd-718f-4812-99d0-85ab356fd9db/volumes" Nov 23 09:07:33 crc kubenswrapper[4906]: I1123 09:07:33.505022 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" event={"ID":"cdc0ed17-e067-446e-9b4f-5027fd193264","Type":"ContainerStarted","Data":"0c595c1a4f122e7620a187e0d38e78ebc8dd65681f422690963ec2176aa81533"} Nov 23 09:07:33 crc kubenswrapper[4906]: I1123 09:07:33.505569 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" event={"ID":"cdc0ed17-e067-446e-9b4f-5027fd193264","Type":"ContainerStarted","Data":"1ec118ee00be6f2302e6d2c8f25aa2a3a9932b6a91bfcff6ef8b15d052039183"} Nov 23 09:07:33 crc kubenswrapper[4906]: I1123 09:07:33.528200 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" podStartSLOduration=1.701786601 podStartE2EDuration="7.528178978s" podCreationTimestamp="2025-11-23 09:07:26 +0000 UTC" firstStartedPulling="2025-11-23 09:07:26.569270385 +0000 UTC m=+8262.082661688" lastFinishedPulling="2025-11-23 09:07:32.395662762 +0000 UTC m=+8267.909054065" observedRunningTime="2025-11-23 09:07:33.525632581 +0000 UTC m=+8269.039023884" watchObservedRunningTime="2025-11-23 09:07:33.528178978 +0000 UTC m=+8269.041570291" Nov 23 09:07:36 crc kubenswrapper[4906]: I1123 09:07:36.357103 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:07:36 crc kubenswrapper[4906]: E1123 09:07:36.357606 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:07:48 crc kubenswrapper[4906]: I1123 09:07:48.356188 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:07:48 crc kubenswrapper[4906]: E1123 09:07:48.356876 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:08:02 crc kubenswrapper[4906]: I1123 09:08:02.357021 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:08:02 crc kubenswrapper[4906]: E1123 09:08:02.358088 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:08:13 crc kubenswrapper[4906]: I1123 09:08:13.357207 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:08:13 crc kubenswrapper[4906]: E1123 09:08:13.358906 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.356647 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:08:26 crc kubenswrapper[4906]: E1123 09:08:26.357443 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.569211 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9_785d6bed-d104-4bdf-85d4-efb0a6db387a/util/0.log" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.733062 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9_785d6bed-d104-4bdf-85d4-efb0a6db387a/pull/0.log" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.735711 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9_785d6bed-d104-4bdf-85d4-efb0a6db387a/util/0.log" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.760449 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9_785d6bed-d104-4bdf-85d4-efb0a6db387a/pull/0.log" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.894458 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9_785d6bed-d104-4bdf-85d4-efb0a6db387a/util/0.log" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.903857 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9_785d6bed-d104-4bdf-85d4-efb0a6db387a/extract/0.log" Nov 23 09:08:26 crc kubenswrapper[4906]: I1123 09:08:26.909555 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287bkvr9_785d6bed-d104-4bdf-85d4-efb0a6db387a/pull/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.069162 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-fkvbv_f7a10077-2d8a-4b42-8683-bfc0f336f7d1/kube-rbac-proxy/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.134453 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-fkvbv_f7a10077-2d8a-4b42-8683-bfc0f336f7d1/manager/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.139309 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-59wgw_150266f3-2d08-4b22-b00e-a04fa8419820/kube-rbac-proxy/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.260502 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-59wgw_150266f3-2d08-4b22-b00e-a04fa8419820/manager/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.302578 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-bgshg_3bc07001-7205-48f6-a348-e62df3c098da/kube-rbac-proxy/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.304640 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-bgshg_3bc07001-7205-48f6-a348-e62df3c098da/manager/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.463798 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-jngn6_e8480261-0abe-4b7d-b741-b76fe8126894/kube-rbac-proxy/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.540046 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-jngn6_e8480261-0abe-4b7d-b741-b76fe8126894/manager/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.633111 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-q6mmt_896605e2-1d27-4bda-a8bf-783a9fbd0294/kube-rbac-proxy/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.650369 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-q6mmt_896605e2-1d27-4bda-a8bf-783a9fbd0294/manager/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.722249 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-vqg2n_bb4b51f3-5f0e-4a41-920c-7a179c1e0032/kube-rbac-proxy/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.841463 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-vqg2n_bb4b51f3-5f0e-4a41-920c-7a179c1e0032/manager/0.log" Nov 23 09:08:27 crc kubenswrapper[4906]: I1123 09:08:27.865462 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-hpwgw_29c9510f-29ed-482f-8b63-377479eff207/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.021531 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-hpwgw_29c9510f-29ed-482f-8b63-377479eff207/manager/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.047322 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-xvpjm_c4eaa84e-ff1f-4a9a-9ffe-56350431eb01/manager/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.062691 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-xvpjm_c4eaa84e-ff1f-4a9a-9ffe-56350431eb01/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.199711 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-8tg5j_3247edc2-e84e-4c6b-9b60-cbfb7a7ec442/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.251823 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-8tg5j_3247edc2-e84e-4c6b-9b60-cbfb7a7ec442/manager/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.347970 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-xm76f_45a30ceb-58e3-4e36-96a0-b64f35fa5d5e/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.404649 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-xm76f_45a30ceb-58e3-4e36-96a0-b64f35fa5d5e/manager/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.475761 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-c9xlv_b9bdc0d6-3a2e-4eda-8c73-870215a2018d/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.608944 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-c9xlv_b9bdc0d6-3a2e-4eda-8c73-870215a2018d/manager/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.612185 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-7d26f_f98f97de-580b-460d-9576-1b496313001f/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.671505 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-7d26f_f98f97de-580b-460d-9576-1b496313001f/manager/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.773425 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-kqzb7_962c14dd-6e7a-4927-ba5b-84384bde4312/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.883331 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-kqzb7_962c14dd-6e7a-4927-ba5b-84384bde4312/manager/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.937880 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-q98m4_cdd2bdf7-90ef-42f0-9395-b4889673d69c/kube-rbac-proxy/0.log" Nov 23 09:08:28 crc kubenswrapper[4906]: I1123 09:08:28.955425 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-q98m4_cdd2bdf7-90ef-42f0-9395-b4889673d69c/manager/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.070903 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd4489m6x_d43ca54d-dde3-4178-bb66-1bf010c5dac8/kube-rbac-proxy/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.112101 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd4489m6x_d43ca54d-dde3-4178-bb66-1bf010c5dac8/manager/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.252588 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-jbgnq_44c3b95a-a75b-4121-967a-28d483eb72a2/kube-rbac-proxy/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.365470 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-v78j2_57a5923c-45c0-4acd-ae5f-3920aa1c532f/kube-rbac-proxy/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.555929 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-v78j2_57a5923c-45c0-4acd-ae5f-3920aa1c532f/operator/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.742956 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-gd2gm_128c4cf7-46dc-4438-a13e-ebb14600b794/kube-rbac-proxy/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.755790 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pb6pq_8013380e-43a3-47f5-bd1c-13d2b9525879/registry-server/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.970939 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-ndc2b_dd441d38-8fef-4ee4-8e92-f32682d524c3/kube-rbac-proxy/0.log" Nov 23 09:08:29 crc kubenswrapper[4906]: I1123 09:08:29.975043 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-gd2gm_128c4cf7-46dc-4438-a13e-ebb14600b794/manager/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.006842 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-ndc2b_dd441d38-8fef-4ee4-8e92-f32682d524c3/manager/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.109879 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-jbgnq_44c3b95a-a75b-4121-967a-28d483eb72a2/manager/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.161879 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-5p2c5_f09be5e8-3996-4aa0-956f-cd5b5cd0d08b/operator/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.179203 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-wtlrm_f050d900-ee68-4bdb-ad1f-3e64bf8db957/kube-rbac-proxy/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.285111 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-wtlrm_f050d900-ee68-4bdb-ad1f-3e64bf8db957/manager/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.359016 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-f4jtr_2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239/kube-rbac-proxy/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.397740 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-f4jtr_2ef8f6b6-9851-4a5e-a24e-5d4ccbcf7239/manager/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.475238 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-s4mqg_c72780f9-1a11-4ddc-bf65-3b1eb726be9e/kube-rbac-proxy/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.548965 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-s4mqg_c72780f9-1a11-4ddc-bf65-3b1eb726be9e/manager/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.561543 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-n728k_4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2/kube-rbac-proxy/0.log" Nov 23 09:08:30 crc kubenswrapper[4906]: I1123 09:08:30.602861 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-n728k_4ff2fa48-6adc-4bac-bf2d-a6a1b87f9ce2/manager/0.log" Nov 23 09:08:41 crc kubenswrapper[4906]: I1123 09:08:41.356428 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:08:41 crc kubenswrapper[4906]: E1123 09:08:41.357100 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:08:45 crc kubenswrapper[4906]: I1123 09:08:45.030274 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bdrqj_29b40410-6b1e-44ce-98a1-f5148a2ad676/control-plane-machine-set-operator/0.log" Nov 23 09:08:45 crc kubenswrapper[4906]: I1123 09:08:45.144091 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qdqtn_11e47284-7e15-4254-93d2-8f2487db8c16/kube-rbac-proxy/0.log" Nov 23 09:08:45 crc kubenswrapper[4906]: I1123 09:08:45.182411 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qdqtn_11e47284-7e15-4254-93d2-8f2487db8c16/machine-api-operator/0.log" Nov 23 09:08:56 crc kubenswrapper[4906]: I1123 09:08:56.357176 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:08:56 crc kubenswrapper[4906]: E1123 09:08:56.357850 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:08:56 crc kubenswrapper[4906]: I1123 09:08:56.906844 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-qbrp4_3db5f0b3-e0b2-4cdf-aae1-39d42a31daa6/cert-manager-controller/0.log" Nov 23 09:08:57 crc kubenswrapper[4906]: I1123 09:08:57.098876 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-d8rz9_b8ec1d86-a3b3-48eb-901d-8ddf495fa97e/cert-manager-cainjector/0.log" Nov 23 09:08:57 crc kubenswrapper[4906]: I1123 09:08:57.161290 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-p4rqc_a8e7a6e3-86d5-4d34-88d7-6fd220435b4e/cert-manager-webhook/0.log" Nov 23 09:09:07 crc kubenswrapper[4906]: I1123 09:09:07.357834 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:09:07 crc kubenswrapper[4906]: E1123 09:09:07.358364 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:09:08 crc kubenswrapper[4906]: I1123 09:09:08.206000 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-qgsjq_acf85f16-cd96-4cc5-b271-16444206d205/nmstate-console-plugin/0.log" Nov 23 09:09:08 crc kubenswrapper[4906]: I1123 09:09:08.362043 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-87hwm_a702c630-d383-4ed0-82d5-07e8e074443a/nmstate-handler/0.log" Nov 23 09:09:08 crc kubenswrapper[4906]: I1123 09:09:08.400393 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-hww9n_b16b1f54-fe15-490f-9f7b-d77ca6708c2e/kube-rbac-proxy/0.log" Nov 23 09:09:08 crc kubenswrapper[4906]: I1123 09:09:08.413716 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-hww9n_b16b1f54-fe15-490f-9f7b-d77ca6708c2e/nmstate-metrics/0.log" Nov 23 09:09:08 crc kubenswrapper[4906]: I1123 09:09:08.571778 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-9qw8j_30760439-eebd-40db-a8fc-7c4c355b6988/nmstate-operator/0.log" Nov 23 09:09:08 crc kubenswrapper[4906]: I1123 09:09:08.631236 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-tglkc_8fff20a6-d9f6-451e-9291-56b3d7c27a4d/nmstate-webhook/0.log" Nov 23 09:09:21 crc kubenswrapper[4906]: I1123 09:09:21.873866 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-vpptg_86aa5159-cbbd-44d3-88e0-cfd0c72e4592/kube-rbac-proxy/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.120833 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-frr-files/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.292503 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-vpptg_86aa5159-cbbd-44d3-88e0-cfd0c72e4592/controller/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.356830 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.357371 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-frr-files/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: E1123 09:09:22.357406 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.357532 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-metrics/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.403667 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-reloader/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.448024 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-reloader/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.613910 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-reloader/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.623716 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-metrics/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.625910 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-metrics/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.647809 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-frr-files/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.812202 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-reloader/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.841243 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-frr-files/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.862256 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/controller/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.863901 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/cp-metrics/0.log" Nov 23 09:09:22 crc kubenswrapper[4906]: I1123 09:09:22.983386 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/frr-metrics/0.log" Nov 23 09:09:23 crc kubenswrapper[4906]: I1123 09:09:23.063087 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/kube-rbac-proxy-frr/0.log" Nov 23 09:09:23 crc kubenswrapper[4906]: I1123 09:09:23.099097 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/kube-rbac-proxy/0.log" Nov 23 09:09:23 crc kubenswrapper[4906]: I1123 09:09:23.183003 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/reloader/0.log" Nov 23 09:09:23 crc kubenswrapper[4906]: I1123 09:09:23.297883 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-m45rz_ad6a4f4d-bf8d-4628-8a89-7c67305e3b47/frr-k8s-webhook-server/0.log" Nov 23 09:09:23 crc kubenswrapper[4906]: I1123 09:09:23.463237 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-67454dfd94-m6rht_f2f26030-1eab-4685-a398-009687188c87/manager/0.log" Nov 23 09:09:23 crc kubenswrapper[4906]: I1123 09:09:23.686883 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5fb958c5d-vnq74_deee94a4-f4d0-46f0-9307-99266349749a/webhook-server/0.log" Nov 23 09:09:23 crc kubenswrapper[4906]: I1123 09:09:23.760388 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hgss7_a0c43812-7474-4ecc-a431-2f79a2441bf2/kube-rbac-proxy/0.log" Nov 23 09:09:24 crc kubenswrapper[4906]: I1123 09:09:24.329028 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hgss7_a0c43812-7474-4ecc-a431-2f79a2441bf2/speaker/0.log" Nov 23 09:09:24 crc kubenswrapper[4906]: I1123 09:09:24.550146 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9jk2v_3bcd4535-f910-4cf2-8260-52d2c2506d7b/frr/0.log" Nov 23 09:09:34 crc kubenswrapper[4906]: I1123 09:09:34.357031 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:09:34 crc kubenswrapper[4906]: E1123 09:09:34.357891 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:09:36 crc kubenswrapper[4906]: I1123 09:09:36.712126 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz_9a038721-ff13-448e-8ad9-59941a69a0f1/util/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.057592 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz_9a038721-ff13-448e-8ad9-59941a69a0f1/util/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.079973 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz_9a038721-ff13-448e-8ad9-59941a69a0f1/pull/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.154564 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz_9a038721-ff13-448e-8ad9-59941a69a0f1/pull/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.312441 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz_9a038721-ff13-448e-8ad9-59941a69a0f1/extract/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.319667 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz_9a038721-ff13-448e-8ad9-59941a69a0f1/util/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.373968 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a9cgpz_9a038721-ff13-448e-8ad9-59941a69a0f1/pull/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.508334 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg_86ca56c2-497b-498d-b69f-5f80f17c1765/util/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.693297 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg_86ca56c2-497b-498d-b69f-5f80f17c1765/pull/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.726865 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg_86ca56c2-497b-498d-b69f-5f80f17c1765/util/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.732496 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg_86ca56c2-497b-498d-b69f-5f80f17c1765/pull/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.920398 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg_86ca56c2-497b-498d-b69f-5f80f17c1765/extract/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.925227 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg_86ca56c2-497b-498d-b69f-5f80f17c1765/util/0.log" Nov 23 09:09:37 crc kubenswrapper[4906]: I1123 09:09:37.963070 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eslcdg_86ca56c2-497b-498d-b69f-5f80f17c1765/pull/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.125182 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dmz9g_ec2ed178-c69d-4492-9564-05c560d68db3/extract-utilities/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.251253 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dmz9g_ec2ed178-c69d-4492-9564-05c560d68db3/extract-utilities/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.273857 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dmz9g_ec2ed178-c69d-4492-9564-05c560d68db3/extract-content/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.276977 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dmz9g_ec2ed178-c69d-4492-9564-05c560d68db3/extract-content/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.445119 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dmz9g_ec2ed178-c69d-4492-9564-05c560d68db3/extract-utilities/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.498606 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dmz9g_ec2ed178-c69d-4492-9564-05c560d68db3/extract-content/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.660280 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jgm8r_eecc9e0d-b884-489e-a67d-b4cf63f931c1/extract-utilities/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.906955 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jgm8r_eecc9e0d-b884-489e-a67d-b4cf63f931c1/extract-utilities/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.919802 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jgm8r_eecc9e0d-b884-489e-a67d-b4cf63f931c1/extract-content/0.log" Nov 23 09:09:38 crc kubenswrapper[4906]: I1123 09:09:38.924045 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jgm8r_eecc9e0d-b884-489e-a67d-b4cf63f931c1/extract-content/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.101613 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jgm8r_eecc9e0d-b884-489e-a67d-b4cf63f931c1/extract-content/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.180451 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jgm8r_eecc9e0d-b884-489e-a67d-b4cf63f931c1/extract-utilities/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.420628 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8_2f94e5f1-58e2-4828-a11d-56daa984b926/util/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.474087 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dmz9g_ec2ed178-c69d-4492-9564-05c560d68db3/registry-server/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.679960 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8_2f94e5f1-58e2-4828-a11d-56daa984b926/util/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.693849 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8_2f94e5f1-58e2-4828-a11d-56daa984b926/pull/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.735970 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8_2f94e5f1-58e2-4828-a11d-56daa984b926/pull/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.876574 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8_2f94e5f1-58e2-4828-a11d-56daa984b926/util/0.log" Nov 23 09:09:39 crc kubenswrapper[4906]: I1123 09:09:39.963327 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8_2f94e5f1-58e2-4828-a11d-56daa984b926/pull/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.011622 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62t6b8_2f94e5f1-58e2-4828-a11d-56daa984b926/extract/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.030715 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jgm8r_eecc9e0d-b884-489e-a67d-b4cf63f931c1/registry-server/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.072541 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sml78_8d5c5e8e-1c86-49f2-8af7-158375dc4ef4/marketplace-operator/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.187930 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lffck_ed0a90cb-fcb6-4119-b71e-dcee407aa7e0/extract-utilities/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.341724 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lffck_ed0a90cb-fcb6-4119-b71e-dcee407aa7e0/extract-content/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.357178 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lffck_ed0a90cb-fcb6-4119-b71e-dcee407aa7e0/extract-utilities/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.378348 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lffck_ed0a90cb-fcb6-4119-b71e-dcee407aa7e0/extract-content/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.532602 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lffck_ed0a90cb-fcb6-4119-b71e-dcee407aa7e0/extract-utilities/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.608937 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lffck_ed0a90cb-fcb6-4119-b71e-dcee407aa7e0/extract-content/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.679768 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zspvs_08bc1a64-82e5-46d0-bea5-e21e40666a88/extract-utilities/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.806403 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zspvs_08bc1a64-82e5-46d0-bea5-e21e40666a88/extract-utilities/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.826751 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zspvs_08bc1a64-82e5-46d0-bea5-e21e40666a88/extract-content/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.840150 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-lffck_ed0a90cb-fcb6-4119-b71e-dcee407aa7e0/registry-server/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.910498 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zspvs_08bc1a64-82e5-46d0-bea5-e21e40666a88/extract-content/0.log" Nov 23 09:09:40 crc kubenswrapper[4906]: I1123 09:09:40.991404 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zspvs_08bc1a64-82e5-46d0-bea5-e21e40666a88/extract-utilities/0.log" Nov 23 09:09:41 crc kubenswrapper[4906]: I1123 09:09:41.022979 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zspvs_08bc1a64-82e5-46d0-bea5-e21e40666a88/extract-content/0.log" Nov 23 09:09:41 crc kubenswrapper[4906]: I1123 09:09:41.989314 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zspvs_08bc1a64-82e5-46d0-bea5-e21e40666a88/registry-server/0.log" Nov 23 09:09:46 crc kubenswrapper[4906]: I1123 09:09:46.356322 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:09:46 crc kubenswrapper[4906]: E1123 09:09:46.356883 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:09:58 crc kubenswrapper[4906]: I1123 09:09:58.356065 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:09:58 crc kubenswrapper[4906]: E1123 09:09:58.357390 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:10:10 crc kubenswrapper[4906]: I1123 09:10:10.357695 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:10:10 crc kubenswrapper[4906]: E1123 09:10:10.358488 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:10:22 crc kubenswrapper[4906]: I1123 09:10:22.356471 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:10:22 crc kubenswrapper[4906]: E1123 09:10:22.357598 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:10:33 crc kubenswrapper[4906]: I1123 09:10:33.357891 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:10:33 crc kubenswrapper[4906]: E1123 09:10:33.358925 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:10:43 crc kubenswrapper[4906]: I1123 09:10:43.893151 4906 generic.go:334] "Generic (PLEG): container finished" podID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerID="1ec118ee00be6f2302e6d2c8f25aa2a3a9932b6a91bfcff6ef8b15d052039183" exitCode=0 Nov 23 09:10:43 crc kubenswrapper[4906]: I1123 09:10:43.893309 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" event={"ID":"cdc0ed17-e067-446e-9b4f-5027fd193264","Type":"ContainerDied","Data":"1ec118ee00be6f2302e6d2c8f25aa2a3a9932b6a91bfcff6ef8b15d052039183"} Nov 23 09:10:43 crc kubenswrapper[4906]: I1123 09:10:43.894795 4906 scope.go:117] "RemoveContainer" containerID="1ec118ee00be6f2302e6d2c8f25aa2a3a9932b6a91bfcff6ef8b15d052039183" Nov 23 09:10:44 crc kubenswrapper[4906]: I1123 09:10:44.378426 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qbkj6_must-gather-nqcgv_cdc0ed17-e067-446e-9b4f-5027fd193264/gather/0.log" Nov 23 09:10:46 crc kubenswrapper[4906]: I1123 09:10:46.356617 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:10:46 crc kubenswrapper[4906]: E1123 09:10:46.356978 4906 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2xgvr_openshift-machine-config-operator(ec5c6569-64d4-4591-bbac-b70ebcca836f)\"" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" Nov 23 09:10:51 crc kubenswrapper[4906]: I1123 09:10:51.660907 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qbkj6/must-gather-nqcgv"] Nov 23 09:10:51 crc kubenswrapper[4906]: I1123 09:10:51.661907 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerName="copy" containerID="cri-o://0c595c1a4f122e7620a187e0d38e78ebc8dd65681f422690963ec2176aa81533" gracePeriod=2 Nov 23 09:10:51 crc kubenswrapper[4906]: I1123 09:10:51.668525 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qbkj6/must-gather-nqcgv"] Nov 23 09:10:51 crc kubenswrapper[4906]: I1123 09:10:51.966527 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qbkj6_must-gather-nqcgv_cdc0ed17-e067-446e-9b4f-5027fd193264/copy/0.log" Nov 23 09:10:51 crc kubenswrapper[4906]: I1123 09:10:51.967402 4906 generic.go:334] "Generic (PLEG): container finished" podID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerID="0c595c1a4f122e7620a187e0d38e78ebc8dd65681f422690963ec2176aa81533" exitCode=143 Nov 23 09:10:51 crc kubenswrapper[4906]: I1123 09:10:51.967480 4906 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe095876145d8b8fd2ab0f8aa9ca0824afceddd5b8c5d6309496245f1ddbd084" Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:51.998994 4906 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qbkj6_must-gather-nqcgv_cdc0ed17-e067-446e-9b4f-5027fd193264/copy/0.log" Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:51.999405 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:52.091711 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwcjh\" (UniqueName: \"kubernetes.io/projected/cdc0ed17-e067-446e-9b4f-5027fd193264-kube-api-access-rwcjh\") pod \"cdc0ed17-e067-446e-9b4f-5027fd193264\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:52.091877 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc0ed17-e067-446e-9b4f-5027fd193264-must-gather-output\") pod \"cdc0ed17-e067-446e-9b4f-5027fd193264\" (UID: \"cdc0ed17-e067-446e-9b4f-5027fd193264\") " Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:52.098098 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc0ed17-e067-446e-9b4f-5027fd193264-kube-api-access-rwcjh" (OuterVolumeSpecName: "kube-api-access-rwcjh") pod "cdc0ed17-e067-446e-9b4f-5027fd193264" (UID: "cdc0ed17-e067-446e-9b4f-5027fd193264"). InnerVolumeSpecName "kube-api-access-rwcjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:52.184470 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc0ed17-e067-446e-9b4f-5027fd193264-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cdc0ed17-e067-446e-9b4f-5027fd193264" (UID: "cdc0ed17-e067-446e-9b4f-5027fd193264"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:52.193400 4906 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc0ed17-e067-446e-9b4f-5027fd193264-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:52.193436 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwcjh\" (UniqueName: \"kubernetes.io/projected/cdc0ed17-e067-446e-9b4f-5027fd193264-kube-api-access-rwcjh\") on node \"crc\" DevicePath \"\"" Nov 23 09:10:52 crc kubenswrapper[4906]: I1123 09:10:52.973820 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qbkj6/must-gather-nqcgv" Nov 23 09:10:53 crc kubenswrapper[4906]: I1123 09:10:53.369369 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" path="/var/lib/kubelet/pods/cdc0ed17-e067-446e-9b4f-5027fd193264/volumes" Nov 23 09:10:59 crc kubenswrapper[4906]: I1123 09:10:59.356268 4906 scope.go:117] "RemoveContainer" containerID="e40768023a8208f3642fc824fc88577f248ca0f224a057c113ead43294f04357" Nov 23 09:11:00 crc kubenswrapper[4906]: I1123 09:11:00.034190 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" event={"ID":"ec5c6569-64d4-4591-bbac-b70ebcca836f","Type":"ContainerStarted","Data":"32eb7e88b189ab6ce6131d405eb0665a42f2cfd79853f4e14de7ddaa960a677c"} Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.529611 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q6cr9"] Nov 23 09:13:14 crc kubenswrapper[4906]: E1123 09:13:14.530756 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="registry-server" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.530778 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="registry-server" Nov 23 09:13:14 crc kubenswrapper[4906]: E1123 09:13:14.530805 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerName="gather" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.530819 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerName="gather" Nov 23 09:13:14 crc kubenswrapper[4906]: E1123 09:13:14.530841 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerName="copy" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.530856 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerName="copy" Nov 23 09:13:14 crc kubenswrapper[4906]: E1123 09:13:14.530873 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="extract-content" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.530887 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="extract-content" Nov 23 09:13:14 crc kubenswrapper[4906]: E1123 09:13:14.530936 4906 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="extract-utilities" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.530949 4906 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="extract-utilities" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.531237 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerName="gather" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.531270 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc0ed17-e067-446e-9b4f-5027fd193264" containerName="copy" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.531318 4906 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93ec0bd-718f-4812-99d0-85ab356fd9db" containerName="registry-server" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.533341 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.547062 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6cr9"] Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.662056 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-catalog-content\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.662171 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-utilities\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.662207 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm9kv\" (UniqueName: \"kubernetes.io/projected/90c13b54-9b82-4079-9e7a-b037c6a30c92-kube-api-access-jm9kv\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.763276 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-catalog-content\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.763352 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-utilities\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.763383 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm9kv\" (UniqueName: \"kubernetes.io/projected/90c13b54-9b82-4079-9e7a-b037c6a30c92-kube-api-access-jm9kv\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.764019 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-catalog-content\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.764083 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-utilities\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.795115 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm9kv\" (UniqueName: \"kubernetes.io/projected/90c13b54-9b82-4079-9e7a-b037c6a30c92-kube-api-access-jm9kv\") pod \"certified-operators-q6cr9\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:14 crc kubenswrapper[4906]: I1123 09:13:14.897972 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:15 crc kubenswrapper[4906]: I1123 09:13:15.420644 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6cr9"] Nov 23 09:13:16 crc kubenswrapper[4906]: I1123 09:13:16.154906 4906 generic.go:334] "Generic (PLEG): container finished" podID="90c13b54-9b82-4079-9e7a-b037c6a30c92" containerID="42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264" exitCode=0 Nov 23 09:13:16 crc kubenswrapper[4906]: I1123 09:13:16.155015 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6cr9" event={"ID":"90c13b54-9b82-4079-9e7a-b037c6a30c92","Type":"ContainerDied","Data":"42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264"} Nov 23 09:13:16 crc kubenswrapper[4906]: I1123 09:13:16.156082 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6cr9" event={"ID":"90c13b54-9b82-4079-9e7a-b037c6a30c92","Type":"ContainerStarted","Data":"4aaf62f831355f0d936c3d0e5e40bdfe5469bc8d36e92439a20d991bec2061cc"} Nov 23 09:13:16 crc kubenswrapper[4906]: I1123 09:13:16.156851 4906 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 09:13:16 crc kubenswrapper[4906]: I1123 09:13:16.919562 4906 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p4v2v"] Nov 23 09:13:16 crc kubenswrapper[4906]: I1123 09:13:16.921455 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:16 crc kubenswrapper[4906]: I1123 09:13:16.932180 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4v2v"] Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.000743 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-catalog-content\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.000855 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-utilities\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.000897 4906 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2ms7\" (UniqueName: \"kubernetes.io/projected/3201ce02-f9a2-4a57-a2fb-9789360e350a-kube-api-access-x2ms7\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.102647 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2ms7\" (UniqueName: \"kubernetes.io/projected/3201ce02-f9a2-4a57-a2fb-9789360e350a-kube-api-access-x2ms7\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.102735 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-catalog-content\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.102816 4906 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-utilities\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.103243 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-utilities\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.103296 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-catalog-content\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.120207 4906 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2ms7\" (UniqueName: \"kubernetes.io/projected/3201ce02-f9a2-4a57-a2fb-9789360e350a-kube-api-access-x2ms7\") pod \"redhat-operators-p4v2v\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.171159 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6cr9" event={"ID":"90c13b54-9b82-4079-9e7a-b037c6a30c92","Type":"ContainerStarted","Data":"04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d"} Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.250917 4906 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:17 crc kubenswrapper[4906]: I1123 09:13:17.757835 4906 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4v2v"] Nov 23 09:13:17 crc kubenswrapper[4906]: W1123 09:13:17.758645 4906 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3201ce02_f9a2_4a57_a2fb_9789360e350a.slice/crio-5fde8c7da97a88404398d2240f056988af462925099e496c405225c682410428 WatchSource:0}: Error finding container 5fde8c7da97a88404398d2240f056988af462925099e496c405225c682410428: Status 404 returned error can't find the container with id 5fde8c7da97a88404398d2240f056988af462925099e496c405225c682410428 Nov 23 09:13:18 crc kubenswrapper[4906]: I1123 09:13:18.178461 4906 generic.go:334] "Generic (PLEG): container finished" podID="3201ce02-f9a2-4a57-a2fb-9789360e350a" containerID="3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b" exitCode=0 Nov 23 09:13:18 crc kubenswrapper[4906]: I1123 09:13:18.178565 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4v2v" event={"ID":"3201ce02-f9a2-4a57-a2fb-9789360e350a","Type":"ContainerDied","Data":"3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b"} Nov 23 09:13:18 crc kubenswrapper[4906]: I1123 09:13:18.178788 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4v2v" event={"ID":"3201ce02-f9a2-4a57-a2fb-9789360e350a","Type":"ContainerStarted","Data":"5fde8c7da97a88404398d2240f056988af462925099e496c405225c682410428"} Nov 23 09:13:18 crc kubenswrapper[4906]: I1123 09:13:18.180777 4906 generic.go:334] "Generic (PLEG): container finished" podID="90c13b54-9b82-4079-9e7a-b037c6a30c92" containerID="04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d" exitCode=0 Nov 23 09:13:18 crc kubenswrapper[4906]: I1123 09:13:18.180820 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6cr9" event={"ID":"90c13b54-9b82-4079-9e7a-b037c6a30c92","Type":"ContainerDied","Data":"04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d"} Nov 23 09:13:19 crc kubenswrapper[4906]: I1123 09:13:19.189533 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4v2v" event={"ID":"3201ce02-f9a2-4a57-a2fb-9789360e350a","Type":"ContainerStarted","Data":"c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6"} Nov 23 09:13:19 crc kubenswrapper[4906]: I1123 09:13:19.191705 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6cr9" event={"ID":"90c13b54-9b82-4079-9e7a-b037c6a30c92","Type":"ContainerStarted","Data":"71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314"} Nov 23 09:13:19 crc kubenswrapper[4906]: I1123 09:13:19.223906 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q6cr9" podStartSLOduration=2.797155616 podStartE2EDuration="5.223881085s" podCreationTimestamp="2025-11-23 09:13:14 +0000 UTC" firstStartedPulling="2025-11-23 09:13:16.15662042 +0000 UTC m=+8611.670011723" lastFinishedPulling="2025-11-23 09:13:18.583345899 +0000 UTC m=+8614.096737192" observedRunningTime="2025-11-23 09:13:19.220618687 +0000 UTC m=+8614.734009990" watchObservedRunningTime="2025-11-23 09:13:19.223881085 +0000 UTC m=+8614.737272418" Nov 23 09:13:20 crc kubenswrapper[4906]: I1123 09:13:20.202159 4906 generic.go:334] "Generic (PLEG): container finished" podID="3201ce02-f9a2-4a57-a2fb-9789360e350a" containerID="c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6" exitCode=0 Nov 23 09:13:20 crc kubenswrapper[4906]: I1123 09:13:20.202295 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4v2v" event={"ID":"3201ce02-f9a2-4a57-a2fb-9789360e350a","Type":"ContainerDied","Data":"c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6"} Nov 23 09:13:20 crc kubenswrapper[4906]: I1123 09:13:20.946214 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:13:20 crc kubenswrapper[4906]: I1123 09:13:20.946512 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 09:13:21 crc kubenswrapper[4906]: I1123 09:13:21.213440 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4v2v" event={"ID":"3201ce02-f9a2-4a57-a2fb-9789360e350a","Type":"ContainerStarted","Data":"64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff"} Nov 23 09:13:21 crc kubenswrapper[4906]: I1123 09:13:21.243037 4906 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p4v2v" podStartSLOduration=2.799656389 podStartE2EDuration="5.243015641s" podCreationTimestamp="2025-11-23 09:13:16 +0000 UTC" firstStartedPulling="2025-11-23 09:13:18.1796388 +0000 UTC m=+8613.693030103" lastFinishedPulling="2025-11-23 09:13:20.622998052 +0000 UTC m=+8616.136389355" observedRunningTime="2025-11-23 09:13:21.234235477 +0000 UTC m=+8616.747626780" watchObservedRunningTime="2025-11-23 09:13:21.243015641 +0000 UTC m=+8616.756406984" Nov 23 09:13:24 crc kubenswrapper[4906]: I1123 09:13:24.898492 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:24 crc kubenswrapper[4906]: I1123 09:13:24.898773 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:24 crc kubenswrapper[4906]: I1123 09:13:24.943849 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:25 crc kubenswrapper[4906]: I1123 09:13:25.309577 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:26 crc kubenswrapper[4906]: I1123 09:13:26.113893 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6cr9"] Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.251125 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.251190 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.256537 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q6cr9" podUID="90c13b54-9b82-4079-9e7a-b037c6a30c92" containerName="registry-server" containerID="cri-o://71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314" gracePeriod=2 Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.298039 4906 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.708609 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.776352 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-catalog-content\") pod \"90c13b54-9b82-4079-9e7a-b037c6a30c92\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.776430 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm9kv\" (UniqueName: \"kubernetes.io/projected/90c13b54-9b82-4079-9e7a-b037c6a30c92-kube-api-access-jm9kv\") pod \"90c13b54-9b82-4079-9e7a-b037c6a30c92\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.776515 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-utilities\") pod \"90c13b54-9b82-4079-9e7a-b037c6a30c92\" (UID: \"90c13b54-9b82-4079-9e7a-b037c6a30c92\") " Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.777395 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-utilities" (OuterVolumeSpecName: "utilities") pod "90c13b54-9b82-4079-9e7a-b037c6a30c92" (UID: "90c13b54-9b82-4079-9e7a-b037c6a30c92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.781679 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90c13b54-9b82-4079-9e7a-b037c6a30c92-kube-api-access-jm9kv" (OuterVolumeSpecName: "kube-api-access-jm9kv") pod "90c13b54-9b82-4079-9e7a-b037c6a30c92" (UID: "90c13b54-9b82-4079-9e7a-b037c6a30c92"). InnerVolumeSpecName "kube-api-access-jm9kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.877778 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm9kv\" (UniqueName: \"kubernetes.io/projected/90c13b54-9b82-4079-9e7a-b037c6a30c92-kube-api-access-jm9kv\") on node \"crc\" DevicePath \"\"" Nov 23 09:13:27 crc kubenswrapper[4906]: I1123 09:13:27.877805 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.265591 4906 generic.go:334] "Generic (PLEG): container finished" podID="90c13b54-9b82-4079-9e7a-b037c6a30c92" containerID="71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314" exitCode=0 Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.265684 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6cr9" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.265799 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6cr9" event={"ID":"90c13b54-9b82-4079-9e7a-b037c6a30c92","Type":"ContainerDied","Data":"71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314"} Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.265831 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6cr9" event={"ID":"90c13b54-9b82-4079-9e7a-b037c6a30c92","Type":"ContainerDied","Data":"4aaf62f831355f0d936c3d0e5e40bdfe5469bc8d36e92439a20d991bec2061cc"} Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.265853 4906 scope.go:117] "RemoveContainer" containerID="71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.284023 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90c13b54-9b82-4079-9e7a-b037c6a30c92" (UID: "90c13b54-9b82-4079-9e7a-b037c6a30c92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.286996 4906 scope.go:117] "RemoveContainer" containerID="04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.310828 4906 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.314904 4906 scope.go:117] "RemoveContainer" containerID="42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.334490 4906 scope.go:117] "RemoveContainer" containerID="71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314" Nov 23 09:13:28 crc kubenswrapper[4906]: E1123 09:13:28.336131 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314\": container with ID starting with 71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314 not found: ID does not exist" containerID="71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.336264 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314"} err="failed to get container status \"71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314\": rpc error: code = NotFound desc = could not find container \"71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314\": container with ID starting with 71506c2058f9a3f16f1dc02a206b47005afcdd984113012a3ed6195e2351b314 not found: ID does not exist" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.336357 4906 scope.go:117] "RemoveContainer" containerID="04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d" Nov 23 09:13:28 crc kubenswrapper[4906]: E1123 09:13:28.336795 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d\": container with ID starting with 04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d not found: ID does not exist" containerID="04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.336857 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d"} err="failed to get container status \"04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d\": rpc error: code = NotFound desc = could not find container \"04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d\": container with ID starting with 04b48be9011fc30805f0e514a44b59feb03c802284e274c5330e4c26d235815d not found: ID does not exist" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.336932 4906 scope.go:117] "RemoveContainer" containerID="42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264" Nov 23 09:13:28 crc kubenswrapper[4906]: E1123 09:13:28.337231 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264\": container with ID starting with 42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264 not found: ID does not exist" containerID="42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.337350 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264"} err="failed to get container status \"42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264\": rpc error: code = NotFound desc = could not find container \"42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264\": container with ID starting with 42363291f2d7b220d6ef022b1d3026976a98399148feecc5be3ac67a982c0264 not found: ID does not exist" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.385008 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90c13b54-9b82-4079-9e7a-b037c6a30c92-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.607192 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6cr9"] Nov 23 09:13:28 crc kubenswrapper[4906]: I1123 09:13:28.617042 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q6cr9"] Nov 23 09:13:29 crc kubenswrapper[4906]: I1123 09:13:29.114439 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4v2v"] Nov 23 09:13:29 crc kubenswrapper[4906]: I1123 09:13:29.365590 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90c13b54-9b82-4079-9e7a-b037c6a30c92" path="/var/lib/kubelet/pods/90c13b54-9b82-4079-9e7a-b037c6a30c92/volumes" Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.278197 4906 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p4v2v" podUID="3201ce02-f9a2-4a57-a2fb-9789360e350a" containerName="registry-server" containerID="cri-o://64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff" gracePeriod=2 Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.649911 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.718180 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-utilities\") pod \"3201ce02-f9a2-4a57-a2fb-9789360e350a\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.718449 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-catalog-content\") pod \"3201ce02-f9a2-4a57-a2fb-9789360e350a\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.718892 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-utilities" (OuterVolumeSpecName: "utilities") pod "3201ce02-f9a2-4a57-a2fb-9789360e350a" (UID: "3201ce02-f9a2-4a57-a2fb-9789360e350a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.718678 4906 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2ms7\" (UniqueName: \"kubernetes.io/projected/3201ce02-f9a2-4a57-a2fb-9789360e350a-kube-api-access-x2ms7\") pod \"3201ce02-f9a2-4a57-a2fb-9789360e350a\" (UID: \"3201ce02-f9a2-4a57-a2fb-9789360e350a\") " Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.719984 4906 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.724891 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3201ce02-f9a2-4a57-a2fb-9789360e350a-kube-api-access-x2ms7" (OuterVolumeSpecName: "kube-api-access-x2ms7") pod "3201ce02-f9a2-4a57-a2fb-9789360e350a" (UID: "3201ce02-f9a2-4a57-a2fb-9789360e350a"). InnerVolumeSpecName "kube-api-access-x2ms7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.800796 4906 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3201ce02-f9a2-4a57-a2fb-9789360e350a" (UID: "3201ce02-f9a2-4a57-a2fb-9789360e350a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.821540 4906 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3201ce02-f9a2-4a57-a2fb-9789360e350a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 09:13:30 crc kubenswrapper[4906]: I1123 09:13:30.821852 4906 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2ms7\" (UniqueName: \"kubernetes.io/projected/3201ce02-f9a2-4a57-a2fb-9789360e350a-kube-api-access-x2ms7\") on node \"crc\" DevicePath \"\"" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.287423 4906 generic.go:334] "Generic (PLEG): container finished" podID="3201ce02-f9a2-4a57-a2fb-9789360e350a" containerID="64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff" exitCode=0 Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.287480 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4v2v" event={"ID":"3201ce02-f9a2-4a57-a2fb-9789360e350a","Type":"ContainerDied","Data":"64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff"} Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.287500 4906 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4v2v" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.287533 4906 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4v2v" event={"ID":"3201ce02-f9a2-4a57-a2fb-9789360e350a","Type":"ContainerDied","Data":"5fde8c7da97a88404398d2240f056988af462925099e496c405225c682410428"} Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.287561 4906 scope.go:117] "RemoveContainer" containerID="64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.303621 4906 scope.go:117] "RemoveContainer" containerID="c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.324032 4906 scope.go:117] "RemoveContainer" containerID="3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.338074 4906 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4v2v"] Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.345283 4906 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p4v2v"] Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.351006 4906 scope.go:117] "RemoveContainer" containerID="64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff" Nov 23 09:13:31 crc kubenswrapper[4906]: E1123 09:13:31.351514 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff\": container with ID starting with 64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff not found: ID does not exist" containerID="64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.351607 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff"} err="failed to get container status \"64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff\": rpc error: code = NotFound desc = could not find container \"64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff\": container with ID starting with 64b9c65b894c26969ca5d1b2ccb15a1dd0c44c3dc3b6b195098e47f7e0236cff not found: ID does not exist" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.351711 4906 scope.go:117] "RemoveContainer" containerID="c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6" Nov 23 09:13:31 crc kubenswrapper[4906]: E1123 09:13:31.352083 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6\": container with ID starting with c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6 not found: ID does not exist" containerID="c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.352122 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6"} err="failed to get container status \"c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6\": rpc error: code = NotFound desc = could not find container \"c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6\": container with ID starting with c8735f3dab99b2dc7b01e5a6748277a36748120824b463c534eaf0d269740fb6 not found: ID does not exist" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.352148 4906 scope.go:117] "RemoveContainer" containerID="3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b" Nov 23 09:13:31 crc kubenswrapper[4906]: E1123 09:13:31.352399 4906 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b\": container with ID starting with 3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b not found: ID does not exist" containerID="3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.352440 4906 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b"} err="failed to get container status \"3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b\": rpc error: code = NotFound desc = could not find container \"3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b\": container with ID starting with 3e01dcc4ff54e9d09d5d9a586ac56baac4fe8b0d91ca19c8a6674cb6bbd0530b not found: ID does not exist" Nov 23 09:13:31 crc kubenswrapper[4906]: I1123 09:13:31.365533 4906 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3201ce02-f9a2-4a57-a2fb-9789360e350a" path="/var/lib/kubelet/pods/3201ce02-f9a2-4a57-a2fb-9789360e350a/volumes" Nov 23 09:13:50 crc kubenswrapper[4906]: I1123 09:13:50.946020 4906 patch_prober.go:28] interesting pod/machine-config-daemon-2xgvr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 09:13:50 crc kubenswrapper[4906]: I1123 09:13:50.946638 4906 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2xgvr" podUID="ec5c6569-64d4-4591-bbac-b70ebcca836f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110550127024442 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110550130017351 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110526306016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110526306015454 5ustar corecore